🤖 AI Summary
Existing diffusion-based data augmentation methods suffer from insufficient diversity guarantees and require scaling the dataset by 10–30× to improve in-distribution performance. To address this, we propose *learning-lag-guided subset augmentation*: selectively applying diffusion-based synthesis only to the subset of samples (30%–40%) that remain poorly classified early in training—i.e., those exhibiting high learning lag. Theoretical analysis—based on a two-layer CNN—and extensive experiments jointly demonstrate that this strategy homogenizes feature learning speeds across samples and mitigates noise amplification, outperforming full-dataset augmentation. Our method is orthogonal to optimization algorithms and compatible with SGD, SAM, and conventional augmentations. On CIFAR-10, CIFAR-100, and TinyImageNet, it yields an average accuracy gain of 2.8%; notably, SGD augmented with our method surpasses SAM in performance. Robust improvements are observed across diverse model architectures and optimizers.
📝 Abstract
Synthetically augmenting training datasets with diffusion models has been an effective strategy for improving generalization of image classifiers. However, existing techniques struggle to ensure the diversity of generation and increase the size of the data by up to 10-30x to improve the in-distribution performance. In this work, we show that synthetically augmenting part of the data that is not learned early in training outperforms augmenting the entire dataset. By analyzing a two-layer CNN, we prove that this strategy improves generalization by promoting homogeneity in feature learning speed without amplifying noise. Our extensive experiments show that by augmenting only 30%-40% of the data, our method boosts the performance by up to 2.8% in a variety of scenarios, including training ResNet, ViT and DenseNet on CIFAR-10, CIFAR-100, and TinyImageNet, with a range of optimizers including SGD and SAM. Notably, our method applied with SGD outperforms the SOTA optimizer, SAM, on CIFAR-100 and TinyImageNet. It can also easily stack with existing weak and strong augmentation strategies to further boost the performance.