🤖 AI Summary
This work investigates the mechanistic impact of skip connections on the number of linear regions in deep neural networks. For piecewise-linear activation networks, we propose an exact modeling framework grounded in tropical algebra, wherein forward propagation is represented as compositions of tropical polynomials, and design an efficient algorithm to analytically compute the total number of linear regions across the entire network. Theoretical analysis and empirical evaluation demonstrate that skip connections substantially increase the number of linear regions, and this growth exhibits structural interpretability: it stems from mitigating gradient degradation and enhancing functional expressivity, thereby improving training stability and suppressing overfitting. To our knowledge, this is the first work to establish a quantitative link between skip connections and the geometric complexity of linear regions—measured via their count—providing novel evidence from tropical geometry for the generalization advantage of residual architectures.
📝 Abstract
Neural networks are important tools in machine learning. Representing piecewise linear activation functions with tropical arithmetic enables the application of tropical geometry. Algorithms are presented to compute regions where the neural networks are linear maps. Through computational experiments, we provide insights on the difficulty to train neural networks, in particular on the problems of overfitting and on the benefits of skip connections.