đ¤ AI Summary
Existing deep neural network training relies heavily on validation sets to detect overfitting/underfitting and tune regularization hyperparametersâe.g., weight decay (WD)âintroducing computational overhead and validation-set dependency. To address this, we propose the OverfittingâUnderfitting Indicator (OUI), defined as the ratio of gradient norm to parameter norm during training. Grounded in gradient analysis and the intrinsic relationship between gradient dynamics and L2 regularization, OUI enables early, stable, and validation-free assessment of generalization behaviorâoften before loss or accuracy convergenceâand supports dynamic monitoring and adaptive WD selection. Experiments on CIFAR-100, TinyImageNet, and ImageNet-1K demonstrate that OUI stabilizes rapidly in early training stages and consistently improves final generalization performance. Across multiple architecturesâincluding ResNet, DenseNet, and EfficientNetâOUI-guided WD optimization reduces hyperparameter search time by over 60% while achieving higher validation accuracy than conventional validation-set-based tuning.
đ Abstract
We introduce the Overfitting-Underfitting Indicator (OUI), a novel tool for monitoring the training dynamics of Deep Neural Networks (DNNs) and identifying optimal regularization hyperparameters. Specifically, we validate that OUI can effectively guide the selection of the Weight Decay (WD) hyperparameter by indicating whether a model is overfitting or underfitting during training without requiring validation data. Through experiments on DenseNet-BC-100 with CIFAR- 100, EfficientNet-B0 with TinyImageNet and ResNet-34 with ImageNet-1K, we show that maintaining OUI within a prescribed interval correlates strongly with improved generalization and validation scores. Notably, OUI converges significantly faster than traditional metrics such as loss or accuracy, enabling practitioners to identify optimal WD (hyperparameter) values within the early stages of training. By leveraging OUI as a reliable indicator, we can determine early in training whether the chosen WD value leads the model to underfit the training data, overfit, or strike a well-balanced trade-off that maximizes validation scores. This enables more precise WD tuning for optimal performance on the tested datasets and DNNs. All code for reproducing these experiments is available at https://github.com/AlbertoFdezHdez/OUI.