π€ AI Summary
This work identifies a previously overlooked mechanism by which data augmentation harms convergence stability during early training of deep neural networks, attributing it to an implicit consistency regularization that induces an expectation risk gap. We provide the first formal proof that the augmented expected risk decomposes as the original risk plus an interpretable consistency regularization gap term. Leveraging functional analysis and statistical learning theory, we establish an artificial shift population risk framework to rigorously characterize the detrimental impact of this gap. Guided by this theoretical analysis, we propose a principled compensation strategy that improves both generalization and training stability. Extensive experiments across standard, out-of-distribution, and long-tailed classification benchmarks demonstrate consistent and significant gains over state-of-the-art baselines. Our implementation is publicly available in PyTorch.
π Abstract
Data augmentation is an important technique in training deep neural networks as it enhances their ability to generalize and remain robust. While data augmentation is commonly used to expand the sample size and act as a consistency regularization term, there is a lack of research on the relationship between them. To address this gap, this paper introduces a more comprehensive mathematical framework for data augmentation. Through this framework, we establish that the expected risk of the shifted population is the sum of the original population risk and a gap term, which can be interpreted as a consistency regularization term. The paper also provides a theoretical understanding of this gap, highlighting its negative effects on the early stages of training. We also propose a method to mitigate these effects. To validate our approach, we conducted experiments using same data augmentation techniques and computing resources under several scenarios, including standard training, out-of-distribution, and imbalanced classification. The results demonstrate that our methods surpass compared methods under all scenarios in terms of generalization ability and convergence stability. We provide our code implementation at the following link: https://github.com/ydlsfhll/ASPR.