Solving Over-Smoothing in GNNs via Nonlocal Message Passing: Algebraic Smoothing and Depth Scalability

📅 2025-12-09
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
The relationship between layer normalization (LN) placement and oversmoothing in graph neural networks (GNNs) remains unclear: Pre-LN mitigates oversmoothing but suffers from the depth curse, whereas Post-LN alleviates the depth curse yet exacerbates oversmoothing. Method: We identify a fundamental trade-off between smoothing control and depth scalability, and propose a Post-LN-based non-local message passing mechanism. Leveraging algebraic smoothing theory, our approach enables controllable information propagation without introducing additional parameters. Contribution/Results: Guided by theoretically grounded normalization dynamics analysis, our method achieves both deep scalability and oversmoothing suppression. Empirical evaluation on five benchmark datasets demonstrates effectiveness—supporting GNNs up to 256 layers—while significantly improving performance and maintaining model efficiency.

Technology Category

Application Category

📝 Abstract
The relationship between Layer Normalization (LN) placement and the over-smoothing phenomenon remains underexplored. We identify a critical dilemma: Pre-LN architectures avoid over-smoothing but suffer from the curse of depth, while Post-LN architectures bypass the curse of depth but experience over-smoothing. To resolve this, we propose a new method based on Post-LN that induces algebraic smoothing, preventing over-smoothing without the curse of depth. Empirical results across five benchmarks demonstrate that our approach supports deeper networks (up to 256 layers) and improves performance, requiring no additional parameters. Key contributions: Theoretical Characterization: Analysis of LN dynamics and their impact on over-smoothing and the curse of depth. A Principled Solution: A parameter-efficient method that induces algebraic smoothing and avoids over-smoothing and the curse of depth. Empirical Validation: Extensive experiments showing the effectiveness of the method in deeper GNNs.
Problem

Research questions and friction points this paper is trying to address.

Addresses over-smoothing in GNNs via nonlocal message passing
Resolves dilemma between over-smoothing and curse of depth
Enables deeper GNNs without extra parameters via algebraic smoothing
Innovation

Methods, ideas, or system contributions that make the work stand out.

Post-LN based method induces algebraic smoothing
Prevents over-smoothing without depth curse
Supports deep networks up to 256 layers
🔎 Similar Papers
No similar papers found.
W
Weiqi Guan
School of Mathematical Sciences, Fudan University, Shanghai 200433, China
Junlin He
Junlin He
The Hong Kong Polytechnic University
urban science