๐ค AI Summary
This work addresses the unresolved question of how neighborhood sampling methods systematically affect large-scale graph neural network (GNN) training. We introduce the first unified analytical framework grounded in Neural Tangent Kernel (NTK) and Gaussian Process (GP) theory to model and characterize the impact of diverse sampling strategies on training dynamics and posterior distributions. Our analysis reveals that, while all sampling schemes converge to the same GP posterior as the sample size tends to infinity, they induce markedly distinct posterior covariance structures under finite-sample regimesโleading to incomparable lower bounds on prediction error. Consequently, no universally optimal sampling strategy exists. This work establishes the first theoretical link between sampling behavior and generalization performance in GNNs, providing a principled, NTK/GP-based foundation for sampling design and analysis.
๐ Abstract
Neighborhood sampling is an important ingredient in the training of large-scale graph neural networks. It suppresses the exponential growth of the neighborhood size across network layers and maintains feasible memory consumption and time costs. While it becomes a standard implementation in practice, its systemic behaviors are less understood. We conduct a theoretical analysis by using the tool of neural tangent kernels, which characterize the (analogous) training dynamics of neural networks based on their infinitely wide counterparts -- Gaussian processes (GPs). We study several established neighborhood sampling approaches and the corresponding posterior GP. With limited samples, the posteriors are all different, although they converge to the same one as the sample size increases. Moreover, the posterior covariance, which lower-bounds the mean squared prediction error, is uncomparable, aligning with observations that no sampling approach dominates.