π€ AI Summary
To address the low training efficiency and high memory overhead of Graph Neural Networks (GNNs) on large-scale graphs, this paper proposes an end-to-end learnable adaptive graph coarsening method. Our approach jointly optimizes GNN parameters and node-merging policies during training, employing differentiable K-means clustering on dynamic node embeddings to achieve task-aware graph simplification. Unlike conventional coarsening methods relying on static structural or feature-based heuristics, ours is the first to enable *train-time learnable*, *heterophily-aware* dynamic coarsening. Experiments on both homophilic and heterophilic graph node classification tasks demonstrate that our method significantly reduces computational and memory costsβby up to 5.3Γβwhile preserving or even improving classification accuracy. Visualization further confirms the downstream-task adaptivity of the learned clustering.
π Abstract
We propose an adaptive graph coarsening method to jointly learn graph neural network (GNN) parameters and merge nodes via K-means clustering during training. As real-world graphs grow larger, processing them directly becomes increasingly challenging and sometimes infeasible. Tailoring algorithms to large-scale data may sacrifice performance, so we instead consider graph reduction to decrease the amount of data used during training. In particular, we propose a method to simultaneously train a GNN and coarsen its graph by partitioning nodes via K-means clustering based on their embeddings. Unlike past graph coarsening works, our approach allows us to merge nodes during training. Not only does this preclude coarsening as a preprocessing step, but our node clusters can adapt to the learning task instead of relying solely on graph connectivity and features. Thus, our method is amenable to scenarios that are challenging for other methods, such as heterophilic data. We validate our approach on both homophilic and heterophilic node classification datasets. We further visualize relationships between node embeddings and their corresponding clusters to illustrate that our coarsened graph adapts to the learning task during training.