Graffe: Graph Representation Learning via Diffusion Probabilistic Models

📅 2025-05-08
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This paper addresses the limited semantic modeling capability in graph representation learning by proposing Graffe, the first self-supervised framework for graphs based on diffusion probabilistic models (DPMs). Methodologically, Graffe jointly optimizes a graph encoder and a conditional diffusion decoder: the encoder extracts compact node- or graph-level representations, which serve as conditioning signals to guide the denoising process; the decoder performs conditional denoising via score matching. Theoretically, we establish for the first time that this denoising objective implicitly maximizes conditional mutual information and derive a computable lower bound thereof. Empirically, Graffe achieves superior node and graph classification performance under linear-probe evaluation, attaining state-of-the-art results on 9 out of 11 real-world benchmarks—systematically demonstrating the effectiveness and promise of diffusion models in graph representation learning.

Technology Category

Application Category

📝 Abstract
Diffusion probabilistic models (DPMs), widely recognized for their potential to generate high-quality samples, tend to go unnoticed in representation learning. While recent progress has highlighted their potential for capturing visual semantics, adapting DPMs to graph representation learning remains in its infancy. In this paper, we introduce Graffe, a self-supervised diffusion model proposed for graph representation learning. It features a graph encoder that distills a source graph into a compact representation, which, in turn, serves as the condition to guide the denoising process of the diffusion decoder. To evaluate the effectiveness of our model, we first explore the theoretical foundations of applying diffusion models to representation learning, proving that the denoising objective implicitly maximizes the conditional mutual information between data and its representation. Specifically, we prove that the negative logarithm of the denoising score matching loss is a tractable lower bound for the conditional mutual information. Empirically, we conduct a series of case studies to validate our theoretical insights. In addition, Graffe delivers competitive results under the linear probing setting on node and graph classification tasks, achieving state-of-the-art performance on 9 of the 11 real-world datasets. These findings indicate that powerful generative models, especially diffusion models, serve as an effective tool for graph representation learning.
Problem

Research questions and friction points this paper is trying to address.

Adapting diffusion models to graph representation learning
Proving denoising maximizes conditional mutual information
Achieving state-of-the-art performance on graph tasks
Innovation

Methods, ideas, or system contributions that make the work stand out.

Self-supervised diffusion model for graph learning
Graph encoder guides diffusion decoder denoising
Denoising loss maximizes conditional mutual information
🔎 Similar Papers
No similar papers found.
Dingshuo Chen
Dingshuo Chen
Institute of Automation, Chinese Academy of Sciences
AI for ScienceGraph LearningData Mining
Shuchen Xue
Shuchen Xue
Academy of Mathematics and Systems Science, Chinese Academy of Sciences
Generative ModelsProbabilistic Machine Learning
Liuji Chen
Liuji Chen
Institute of Automation, Chinese Academy of Sciences
LLM AgentTrustworthy AI
Yingheng Wang
Yingheng Wang
Cornell University
Computer Science
Q
Q. Liu
State Key Laboratory of Multimodal Artificial Intelligence Systems (MAIS), Institute of Automation, Chinese Academy of Sciences (CASIA), Beijing 100190, China
S
Shu Wu
State Key Laboratory of Multimodal Artificial Intelligence Systems (MAIS), Institute of Automation, Chinese Academy of Sciences (CASIA), Beijing 100190, China
Z
Zhi-Ming Ma
Academy of Mathematics and Systems Science, Chinese Academy of Sciences (CAS), Beijing 100190, China
L
Liang Wang
State Key Laboratory of Multimodal Artificial Intelligence Systems (MAIS), Institute of Automation, Chinese Academy of Sciences (CASIA), Beijing 100190, China