Shared-Weights Extender and Gradient Voting for Neural Network Expansion

📅 2025-09-23
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Neural networks often suffer from neuron deactivation when dynamically expanded, leading to ineffective capacity growth. To address this, we propose the Shared-Weight Expander (SWE) and the Steepest-Gradient Voting Allocator (SVoD). SWE enforces parameter smoothness and inheritance by coupling newly added neurons with existing ones via weight sharing. SVoD dynamically allocates expansion budgets across layers based on gradient magnitude, enabling hierarchical, adaptive expansion in deep networks. Together, SWE and SVoD support end-to-end backpropagation training. Experiments on four benchmark datasets demonstrate that our method significantly mitigates neuron deactivation. Compared to state-of-the-art expansion strategies and baseline models, it consistently improves both accuracy and parameter efficiency. These results validate the architectural scalability and training stability of the proposed approach.

Technology Category

Application Category

📝 Abstract
Expanding neural networks during training is a promising way to augment capacity without retraining larger models from scratch. However, newly added neurons often fail to adjust to a trained network and become inactive, providing no contribution to capacity growth. We propose the Shared-Weights Extender (SWE), a novel method explicitly designed to prevent inactivity of new neurons by coupling them with existing ones for smooth integration. In parallel, we introduce the Steepest Voting Distributor (SVoD), a gradient-based method for allocating neurons across layers during deep network expansion. Our extensive benchmarking on four datasets shows that our method can effectively suppress neuron inactivity and achieve better performance compared to other expanding methods and baselines.
Problem

Research questions and friction points this paper is trying to address.

Preventing newly added neurons from becoming inactive during network expansion
Allocating neurons effectively across layers in deep network expansion
Augmenting neural network capacity without retraining from scratch
Innovation

Methods, ideas, or system contributions that make the work stand out.

Shared-Weights Extender prevents new neuron inactivity
Steepest Voting Distributor allocates neurons using gradients
Method couples new neurons with existing ones for integration
🔎 Similar Papers
No similar papers found.
N
Nikolas Chatzis
Robotics Institute, Athena Research Center, 15125 Maroussi, Greece; HERON - Hellenic Robotics Center of Excellence, Athens, Greece; School of ECE, National Technical University of Athens, Greece
Ioannis Kordonis
Ioannis Kordonis
National Technical University of Athens
Game TheoryStochastic Control
M
Manos Theodosis
School of Engineering and Applied Sciences Harvard University Cambridge, MA 02138
Petros Maragos
Petros Maragos
Professor of Electrical and Computer Engineering, National Technical University of Athens
computer visionsignal processingspeech&languagemachine learningrobotics