🤖 AI Summary
Neural networks often suffer from neuron deactivation when dynamically expanded, leading to ineffective capacity growth. To address this, we propose the Shared-Weight Expander (SWE) and the Steepest-Gradient Voting Allocator (SVoD). SWE enforces parameter smoothness and inheritance by coupling newly added neurons with existing ones via weight sharing. SVoD dynamically allocates expansion budgets across layers based on gradient magnitude, enabling hierarchical, adaptive expansion in deep networks. Together, SWE and SVoD support end-to-end backpropagation training. Experiments on four benchmark datasets demonstrate that our method significantly mitigates neuron deactivation. Compared to state-of-the-art expansion strategies and baseline models, it consistently improves both accuracy and parameter efficiency. These results validate the architectural scalability and training stability of the proposed approach.
📝 Abstract
Expanding neural networks during training is a promising way to augment capacity without retraining larger models from scratch. However, newly added neurons often fail to adjust to a trained network and become inactive, providing no contribution to capacity growth. We propose the Shared-Weights Extender (SWE), a novel method explicitly designed to prevent inactivity of new neurons by coupling them with existing ones for smooth integration. In parallel, we introduce the Steepest Voting Distributor (SVoD), a gradient-based method for allocating neurons across layers during deep network expansion. Our extensive benchmarking on four datasets shows that our method can effectively suppress neuron inactivity and achieve better performance compared to other expanding methods and baselines.