An Extensive Study on D2C: Overfitting Remediation in Deep Learning Using a Decentralized Approach

📅 2024-11-24
🏛️ arXiv.org
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Deep learning models suffer from overfitting under noisy data, outliers, and small-sample regimes. Existing Divide2Conquer (D2C) methods aggregate sub-models via uniform or heuristic weighting, ignoring their heterogeneous generalization capabilities. Method: We propose an enhanced D2C framework: (i) decentralized data partitioning to construct heterogeneous subsets; (ii) training homogeneous sub-models; and (iii) a generalization-aware dynamic parameter aggregation mechanism—requiring no explicit regularization or data augmentation. Contributions/Results: We theoretically establish convergence and uncover generalization mechanisms via decision-boundary analysis and loss modeling. Extensive experiments on multiple benchmark datasets demonstrate significant improvements over baselines, especially with large-scale data where generalization gains are most pronounced. Our method is orthogonal to mainstream anti-overfitting techniques and enables seamless integration. The implementation is publicly available.

Technology Category

Application Category

📝 Abstract
Overfitting remains a significant challenge in deep learning, often arising from data outliers, noise, and limited training data. To address this, we propose Divide2Conquer (D2C), a novel technique to mitigate overfitting. D2C partitions the training data into multiple subsets and trains identical models independently on each subset. To balance model generalization and subset-specific learning, the model parameters are periodically aggregated and averaged during training. This process enables the learning of robust patterns while minimizing the influence of outliers and noise. Empirical evaluations on benchmark datasets across diverse deep-learning tasks demonstrate that D2C significantly enhances generalization performance, particularly with larger datasets. Our analysis includes evaluations of decision boundaries, loss curves, and other performance metrics, highlighting D2C's effectiveness both as a standalone technique and in combination with other overfitting reduction methods. We further provide a rigorous mathematical justification for D2C's underlying principles and examine its applicability across multiple domains. Finally, we explore the trade-offs associated with D2C and propose strategies to address them, offering a holistic view of its strengths and limitations. This study establishes D2C as a versatile and effective approach to combating overfitting in deep learning. Our codes are publicly available at: https://github.com/Saiful185/Divide2Conquer.
Problem

Research questions and friction points this paper is trying to address.

Addressing overfitting caused by data outliers and limited training data
Improving aggregation of subset models by considering their generalization capabilities
Dynamically weighting model contributions based on validation performance and uncertainty
Innovation

Methods, ideas, or system contributions that make the work stand out.

Dynamic weighting of subset models using validation performance
Aggregation based on both accuracy and prediction uncertainty
Improves generalization by learning from confident edge models
🔎 Similar Papers
No similar papers found.
M
Md. Saiful Bari Siddiqui
Department of Computer Science and Engineering, Brac University, Dhaka, Bangladesh
Md Mohaiminul Islam
Md Mohaiminul Islam
Graduate Student, UNC Chapel Hill
Machine learningComputer VisionMulti-modal deep learningVideo understanding
M
Md. Golam Rabiul Alam
Department of Computer Science and Engineering, Brac University, Dhaka, Bangladesh