Beyond Overfitting: Doubly Adaptive Dropout for Generalizable AU Detection

📅 2025-03-12
🏛️ IEEE Transactions on Affective Computing
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address overfitting and poor generalization in cross-domain facial action unit (AU) detection, this paper proposes a dual-adaptive Dropout mechanism that jointly suppresses domain-specific noise at both the convolutional feature map and Transformer spatial token levels. We introduce a novel Channel Drop Unit (CDU) and Token Drop Unit (TDU) architecture, coupled with layer-wise auxiliary domain classifiers to enable selective dropout of domain-sensitive features. A progressive regularization training strategy is further adopted to balance robustness and semantic information preservation. Our method achieves significant improvements over state-of-the-art approaches across multiple cross-domain AU benchmarks. Attention visualizations clearly localize activation regions for both single and compound AUs, demonstrating strong model interpretability and cross-domain generalization capability.

Technology Category

Application Category

📝 Abstract
Facial Action Units (AUs) are essential for conveying psychological states and emotional expressions. While automatic AU detection systems leveraging deep learning have progressed, they often overfit to specific datasets and individual features, limiting their cross-domain applicability. To overcome these limitations, we propose a doubly adaptive dropout approach for cross-domain AU detection, which enhances the robustness of convolutional feature maps and spatial tokens against domain shifts. This approach includes a Channel Drop Unit (CD-Unit) and a Token Drop Unit (TD-Unit), which work together to reduce domain-specific noise at both the channel and token levels. The CD-Unit preserves domain-agnostic local patterns in feature maps, while the TD-Unit helps the model identify AU relationships generalizable across domains. An auxiliary domain classifier, integrated at each layer, guides the selective omission of domain-sensitive features. To prevent excessive feature dropout, a progressive training strategy is used, allowing for selective exclusion of sensitive features at any model layer. Our method consistently outperforms existing techniques in cross-domain AU detection, as demonstrated by extensive experimental evaluations. Visualizations of attention maps also highlight clear and meaningful patterns related to both individual and combined AUs, further validating the approach's effectiveness.
Problem

Research questions and friction points this paper is trying to address.

Overfitting in AU detection systems limits cross-domain applicability.
Proposed doubly adaptive dropout enhances robustness against domain shifts.
Method outperforms existing techniques in cross-domain AU detection.
Innovation

Methods, ideas, or system contributions that make the work stand out.

Doubly adaptive dropout enhances cross-domain robustness.
Channel and Token Drop Units reduce domain-specific noise.
Progressive training strategy selectively excludes sensitive features.
🔎 Similar Papers
No similar papers found.
Y
Yong Li
School of Computer Science and Engineering and the Key Laboratory of New Generation Artificial Intelligence Technology and Its Interdisciplinary Applications, Southeast University, Nanjing 210096, China
Y
Yi Ren
School of Computer Science and Engineering, Nanjing University of Science and Technology, Nanjing, 210094 China
Xuesong Niu
Xuesong Niu
Institute of Computing Technology; Kuaishou Technology
Affective ComputingComputer Vision
Y
Yi Ding
School of Computer Science and Engineering, Nanyang Technological University, 50 Nanyang Avenue, Singapore, 639798
Xiu-Shen Wei
Xiu-Shen Wei
Professor, Southeast University
Computer VisionMachine LearningArtificial Intelligence
Cuntai Guan
Cuntai Guan
President's Chair Professor, CCDS, Nanyang Technological University
Brain-Computer InterfaceBrain-Computer InterfacesMachine LearningArtificial Intelligence