🤖 AI Summary
Multilayer perceptrons (MLPs) suffer from limited generalization due to high variance in empirical loss estimates. Method: This paper introduces ranked set sampling (RSS)—a novel data resampling strategy for neural network training—to construct ordered data structures that reduce empirical loss variance. We theoretically establish that RSS yields lower variance than simple random sampling and enhances both diversity and consistency among base learners. Furthermore, we propose two variance-aware fusion strategies—integrating empirical exponential loss and logistic loss—to improve loss estimation. Results: Extensive experiments across 12 benchmark datasets demonstrate that RSS-MLP consistently reduces empirical loss variance and achieves stable generalization improvements under both loss functions, empirically validating the critical role of variance control in enhancing MLP generalization.
📝 Abstract
Multilayer perceptron (MLP), one of the most fundamental neural networks, is extensively utilized for classification and regression tasks. In this paper, we establish a new generalization error bound, which reveals how the variance of empirical loss influences the generalization ability of the learning model. Inspired by this learning bound, we advocate to reduce the variance of empirical loss to enhance the ability of MLP. As is well-known, bagging is a popular ensemble method to realize variance reduction. However, bagging produces the base training data sets by the Simple Random Sampling (SRS) method, which exhibits a high degree of randomness. To handle this issue, we introduce an ordered structure in the training data set by Rank Set Sampling (RSS) to further reduce the variance of loss and develop a RSS-MLP method. Theoretical results show that the variance of empirical exponential loss and the logistic loss estimated by RSS are smaller than those estimated by SRS, respectively. To validate the performance of RSS-MLP, we conduct comparison experiments on twelve benchmark data sets in terms of the two convex loss functions under two fusion methods. Extensive experimental results and analysis illustrate the effectiveness and rationality of the propose method.