🤖 AI Summary
Long-range information propagation in Graph Neural Networks (GNNs) is fundamentally hindered by the *oversquashing* problem, causing exponential decay of information flow with increasing hop count. To address this, we introduce SWAN—a novel GNN architecture grounded in dynamical systems theory—that jointly enforces global and local non-dissipativity via dual anti-symmetric parameterizations in both the spatial and weight domains. By ensuring energy conservation and nonlinear stability throughout message passing, SWAN theoretically guarantees constant-rate long-range information propagation. Empirically, SWAN significantly alleviates oversquashing on synthetic and real-world long-range interaction benchmarks. It consistently improves performance on node classification and graph-level prediction tasks, demonstrating enhanced modeling of long-range dependencies. Our work establishes non-dissipative dynamics as a fundamental mechanism for boosting the expressive power of GNNs.
📝 Abstract
A common problem in Message-Passing Neural Networks is oversquashing -- the limited ability to facilitate effective information flow between distant nodes. Oversquashing is attributed to the exponential decay in information transmission as node distances increase. This paper introduces a novel perspective to address oversquashing, leveraging dynamical systems properties of global and local non-dissipativity, that enable the maintenance of a constant information flow rate. We present SWAN, a uniquely parameterized GNN model with antisymmetry both in space and weight domains, as a means to obtain non-dissipativity. Our theoretical analysis asserts that by implementing these properties, SWAN offers an enhanced ability to transmit information over extended distances. Empirical evaluations on synthetic and real-world benchmarks that emphasize long-range interactions validate the theoretical understanding of SWAN, and its ability to mitigate oversquashing.