🤖 AI Summary
This work addresses the counterintuitive phenomenon that Graph Neural Network (GNN) generalization improves with increasing graph size. We establish a statistical generalization theory grounded in the manifold hypothesis—specifically, assuming graph data are sampled from a spectral-domain manifold—to overcome the degradation of conventional generalization bounds with growing node count. Methodologically, we prove for the first time that GNN generalization error decays linearly with graph size on a logarithmic scale, identifying the spectral continuity constant as the key limiting factor; our framework unifies generalization analysis for both node-level and graph-level tasks. Integrating spectral graph theory, Rademacher complexity analysis, and spectral continuity modeling of graph filters, we empirically validate bound tightness across multiple synthetic and real-world graph benchmarks. Our results provide interpretable, theory-driven guidance for GNN architecture design—particularly regarding explicit smoothness constraints on spectral filters.
📝 Abstract
Graph Neural Networks (GNNs) extend convolutional neural networks to operate on graphs. Despite their impressive performances in various graph learning tasks, the theoretical understanding of their generalization capability is still lacking. Previous GNN generalization bounds ignore the underlying graph structures, often leading to bounds that increase with the number of nodes -- a behavior contrary to the one experienced in practice. In this paper, we take a manifold perspective to establish the statistical generalization theory of GNNs on graphs sampled from a manifold in the spectral domain. As demonstrated empirically, we prove that the generalization bounds of GNNs decrease linearly with the size of the graphs in the logarithmic scale, and increase linearly with the spectral continuity constants of the filter functions. Notably, our theory explains both node-level and graph-level tasks. Our result has two implications: i) guaranteeing the generalization of GNNs to unseen data over manifolds; ii) providing insights into the practical design of GNNs, i.e., restrictions on the discriminability of GNNs are necessary to obtain a better generalization performance. We demonstrate our generalization bounds of GNNs using synthetic and multiple real-world datasets.