๐ค AI Summary
This work addresses the inefficiency of manual neural architecture design in resource-constrained telecommunications scenarios. We propose TabGNS, a gradient-driven neural architecture search method tailored for tabular data, which introducesโ for the first timeโa gradient-guided gated neuron selection mechanism to automatically construct compact and efficient models. By leveraging gradient information to guide architectural decisions, TabGNS significantly improves search efficiency and model compression. Empirical evaluations on multiple telecommunications and general-purpose tabular datasets demonstrate that the method reduces model size by 51%โ82% and accelerates the search process by up to 36ร, all while maintaining or even enhancing predictive performance compared to existing approaches.
๐ Abstract
The telecommunications industry is experiencing rapid growth in adopting deep learning for critical tasks such as traffic prediction, signal strength prediction, and quality of service optimisation. However, designing neural network architectures for these applications remains challenging and time-consuming, particularly when targeting compact models suitable for resource-constrained network environments. Therefore, there is a need for automating the model design process to create high-performing models efficiently. This paper introduces TabGNS (Tabular Gated Neuron Selection), a novel gradient-based Neural Architecture Search (NAS) method specifically tailored for tabular data in telecommunications networks. We evaluate TabGNS across multiple telecommunications and generic tabular datasets, demonstrating improvements in prediction performance while reducing the architecture size by 51-82% and reducing the search time by up to 36x compared to state-of-the-art tabular NAS methods. Integrating TabGNS into the model lifecycle management enables automated design of neural networks throughout the lifecycle, accelerating deployment of ML solutions in telecommunications networks.