Single Domain Generalization with Model-aware Parametric Batch-wise Mixup

📅 2025-02-22
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Single-domain generalization (SDG) faces the core challenge of severe distribution shift between the source domain and unseen target domains, leading to poor generalization. To address this, we propose Model-aware Parameterized Batch-wise Mixing (MPBM), a novel data augmentation framework that integrates adversarial query guidance, attention-driven generative networks, and Stochastic Gradient Langevin Dynamics (SGLD) into a controllable synthesis mechanism—enabling generation of semantically consistent and distribution-aligned augmented samples. MPBM extends standard Mixup via parameterized interpolation, broadening the representational coverage of original data and mitigating inter-domain mismatch. Evaluated on multiple standard SDG benchmarks, MPBM consistently outperforms existing methods, achieving state-of-the-art performance. It significantly enhances model robustness and generalization capability on unseen target domains, demonstrating strong practical efficacy for real-world deployment.

Technology Category

Application Category

📝 Abstract
Single Domain Generalization (SDG) remains a formidable challenge in the field of machine learning, particularly when models are deployed in environments that differ significantly from their training domains. In this paper, we propose a novel data augmentation approach, named as Model-aware Parametric Batch-wise Mixup (MPBM), to tackle the challenge of SDG. MPBM deploys adversarial queries generated with stochastic gradient Langevin dynamics, and produces model-aware augmenting instances with a parametric batch-wise mixup generator network that is carefully designed through an innovative attention mechanism. By exploiting inter-feature correlations, the parameterized mixup generator introduces additional versatility in combining features across a batch of instances, thereby enhancing the capacity to generate highly adaptive and informative synthetic instances for specific queries. The synthetic data produced by this adaptable generator network, guided by informative queries, is expected to significantly enrich the representation space covered by the original training dataset and subsequently enhance the prediction model's generalizability across diverse and previously unseen domains. To prevent excessive deviation from the training data, we further incorporate a real-data alignment-based adversarial loss into the learning process of MPBM, regularizing any tendencies toward undesirable expansions. We conduct extensive experiments on several benchmark datasets. The empirical results demonstrate that by augmenting the training set with informative synthesis data, our proposed MPBM method achieves the state-of-the-art performance for single domain generalization.
Problem

Research questions and friction points this paper is trying to address.

Single Domain Generalization challenge
Model-aware Parametric Batch-wise Mixup
Enhancing model generalizability across domains
Innovation

Methods, ideas, or system contributions that make the work stand out.

Parametric Batch-wise Mixup
Adversarial query generation
Real-data alignment loss
🔎 Similar Papers
No similar papers found.