🤖 AI Summary
Existing GNN-to-MLP (G2M) distillation methods lack dynamic trade-off capability between inference cost and accuracy, limiting adaptability to resource-constrained and heterogeneous deployment environments. This paper proposes the first on-demand adjustable G2M knowledge distillation framework that enables flexible precision-efficiency balancing. Its core contributions are: (1) a progressive knowledge distillation mechanism that transfers GNN knowledge in stages; (2) a multi-stage scalable MLP student architecture supporting configurable inference overhead; and (3) a progressive hybrid augmentation strategy integrating structural and feature-level enhancements to improve generalization. Extensive experiments across eight real-world graph datasets demonstrate that our method achieves substantial efficiency gains—average inference latency reduced by 62% and parameter count by 89%—while maintaining controlled accuracy degradation (average drop of only 1.3%). The framework thus effectively supports heterogeneous edge deployment with adaptive precision-efficiency tuning.
📝 Abstract
GNN-to-MLP (G2M) methods have emerged as a promising approach to accelerate Graph Neural Networks (GNNs) by distilling their knowledge into simpler Multi-Layer Perceptrons (MLPs). These methods bridge the gap between the expressive power of GNNs and the computational efficiency of MLPs, making them well-suited for resource-constrained environments. However, existing G2M methods are limited by their inability to flexibly adjust inference cost and accuracy dynamically, a critical requirement for real-world applications where computational resources and time constraints can vary significantly. To address this, we introduce a Progressive framework designed to offer flexible and on-demand trade-offs between inference cost and accuracy for GNN-to-MLP knowledge distillation (ProGMLP). ProGMLP employs a Progressive Training Structure (PTS), where multiple MLP students are trained in sequence, each building on the previous one. Furthermore, ProGMLP incorporates Progressive Knowledge Distillation (PKD) to iteratively refine the distillation process from GNNs to MLPs, and Progressive Mixup Augmentation (PMA) to enhance generalization by progressively generating harder mixed samples. Our approach is validated through comprehensive experiments on eight real-world graph datasets, demonstrating that ProGMLP maintains high accuracy while dynamically adapting to varying runtime scenarios, making it highly effective for deployment in diverse application settings.