🤖 AI Summary
This work addresses the lack of theoretical foundations for residual connections by proposing KITINet—the first neural architecture that models feature propagation as a nonequilibrium particle kinetic process. Methodologically, it discretizes the Boltzmann transport equation to enable physics-informed feature evolution and introduces a channel-wise sparsification mechanism wherein parameters spontaneously condense into dominant channels, enhancing representational efficiency. Innovatively, it is the first to integrate kinetic theory into neural network design, unifying PDE-based numerical solving, stochastic particle system simulation, and adaptive refinement. KITINet achieves state-of-the-art performance across diverse benchmarks—including PDE operator learning, image classification (CIFAR-10/100), and text classification (IMDb/SNLI)—while maintaining near-identical FLOPs compared to classical baselines.
📝 Abstract
Despite the widely recognized success of residual connections in modern neural networks, their design principles remain largely heuristic. This paper introduces KITINet (Kinetics Theory Inspired Network), a novel architecture that reinterprets feature propagation through the lens of non-equilibrium particle dynamics and partial differential equation (PDE) simulation. At its core, we propose a residual module that models feature updates as the stochastic evolution of a particle system, numerically simulated via a discretized solver for the Boltzmann transport equation (BTE). This formulation mimics particle collisions and energy exchange, enabling adaptive feature refinement via physics-informed interactions. Additionally, we reveal that this mechanism induces network parameter condensation during training, where parameters progressively concentrate into a sparse subset of dominant channels. Experiments on scientific computation (PDE operator), image classification (CIFAR-10/100), and text classification (IMDb/SNLI) show consistent improvements over classic network baselines, with negligible increase of FLOPs.