From Cluster Assumption to Graph Convolution: Graph-Based Semi-Supervised Learning Revisited

📅 2023-09-24
🏛️ IEEE Transactions on Neural Networks and Learning Systems
📈 Citations: 6
Influential: 0
📄 PDF
🤖 AI Summary
This paper addresses the fundamental over-smoothing problem in Graph Convolutional Networks (GCNs), arising from the difficulty of jointly modeling graph structure and label information across layers. We theoretically identify its root cause within a unified optimization framework and clarify the essential divergence between conventional graph-based semi-supervised learning (grounded in the cluster assumption) and GCNs in their optimization objectives. Building on this analysis, we propose three novel graph convolution paradigms: (i) supervised OGC, (ii) learning-free structure-preserving GGC, and (iii) multi-scale GGCM—each explicitly unifying label guidance with structural preservation. Experiments demonstrate that our methods significantly mitigate over-smoothing and consistently outperform mainstream models—including GCN and GAT—on benchmark datasets such as Cora and Citeseer. This validates the critical importance of co-modeling structural fidelity and label supervision.
📝 Abstract
Graph-based semi-supervised learning (GSSL) has long been a research focus. Traditional methods are generally shallow learners, based on the cluster assumption. Recently, graph convolutional networks (GCNs) have become the predominant techniques for their promising performance. However, a critical question remains largely unanswered: why do deep GCNs encounter the oversmoothing problem, while traditional shallow GSSL methods do not, despite both progressing through the graph in a similar iterative manner? In this article, we theoretically discuss the relationship between these two types of methods in a unified optimization framework. One of the most intriguing findings is that, unlike traditional ones, typical GCNs may not effectively incorporate both graph structure and label information at each layer. Motivated by this, we propose three simple but powerful graph convolution methods. The first, optimized simple graph convolution (OGC), is a supervised method, which guides the graph convolution process with labels. The others are two “no-learning” unsupervised methods: graph structure preserving graph convolution (GGC) and its multiscale version GGCM, both aiming to preserve the graph structure information during the convolution process. Finally, we conduct extensive experiments to show the effectiveness of our methods.
Problem

Research questions and friction points this paper is trying to address.

Exploring relationship between traditional and GCN-based semi-supervised learning
Addressing GCNs' neglect of graph structure and label information
Proposing improved graph convolution methods for better performance
Innovation

Methods, ideas, or system contributions that make the work stand out.

Proposes supervised OGC method using labels
Introduces unsupervised GGC preserving graph structure
Develops multi-scale GGCM for enhanced performance
🔎 Similar Papers
No similar papers found.
Z
Zheng Wang
Shanghai Jiao Tong University
H
H. Ding
NIO Technology
L
L. Pan
Shanghai Jiao Tong University
J
Jianhua Li
Shanghai Jiao Tong University
Zhiguo Gong
Zhiguo Gong
University of Macau
Philip S. Yu
Philip S. Yu
Professor of Computer Science, University of Illinons at Chicago
Data miningDatabasePrivacy