Towards Graph Foundation Models: A Survey and Beyond

📅 2023-10-18
🏛️ arXiv.org
📈 Citations: 78
Influential: 7
📄 PDF
🤖 AI Summary
Graph Foundation Models (GFMs) lack a formal definition and systematic analysis in graph machine learning. Method: This paper formally defines GFMs’ core characteristics, proposes an evolutionary framework integrating pretraining–adaptation paradigms, and establishes a three-dimensional taxonomy grounded in GNN capability, LLM dependency, and cross-modal synergy. It systematically surveys over 100 state-of-the-art works to distill key technical pathways and constructs the first theoretical framework for GFMs. Contribution/Results: The work identifies six critical challenges—scalable pretraining, unified evaluation benchmarks, among others—and charts corresponding research directions. By rigorously formalizing GFMs and providing a comprehensive analytical foundation, this study fills a foundational definitional gap in the emerging field, enabling principled theoretical development and empirical investigation of graph foundation models.
📝 Abstract
Foundation models have emerged as critical components in a variety of artificial intelligence applications, and showcase significant success in natural language processing and several other domains. Meanwhile, the field of graph machine learning is witnessing a paradigm transition from shallow methods to more sophisticated deep learning approaches. The capabilities of foundation models to generalize and adapt motivate graph machine learning researchers to discuss the potential of developing a new graph learning paradigm. This paradigm envisions models that are pre-trained on extensive graph data and can be adapted for various graph tasks. Despite this burgeoning interest, there is a noticeable lack of clear definitions and systematic analyses pertaining to this new domain. To this end, this article introduces the concept of Graph Foundation Models (GFMs), and offers an exhaustive explanation of their key characteristics and underlying technologies. We proceed to classify the existing work related to GFMs into three distinct categories, based on their dependence on graph neural networks and large language models. In addition to providing a thorough review of the current state of GFMs, this article also outlooks potential avenues for future research in this rapidly evolving domain.
Problem

Research questions and friction points this paper is trying to address.

Explores potential of Graph Foundation Models (GFMs) for graph machine learning.
Lacks clear definitions and systematic analyses for GFMs in current research.
Classifies GFMs based on graph neural networks and large language models.
Innovation

Methods, ideas, or system contributions that make the work stand out.

Introduces Graph Foundation Models (GFMs)
Pre-trained models for diverse graph tasks
Classifies GFMs by GNN and LLM dependencies
🔎 Similar Papers
No similar papers found.
J
Jiawei Liu
School of Computer Science, Beijing University of Posts and Telecommunications, Beijing, China
C
Cheng Yang
School of Computer Science, Beijing University of Posts and Telecommunications, Beijing, China
Z
Zhiyuan Lu
School of Computer Science, Beijing University of Posts and Telecommunications, Beijing, China
J
Junze Chen
School of Computer Science, Beijing University of Posts and Telecommunications, Beijing, China
Yibo Li
Yibo Li
National University of Singapore
LLM
Mengmei Zhang
Mengmei Zhang
China Telecom Bestpay, Beijing, China
Foundation ModelsGNNLarge Language ModelsAdversarial Robustness
T
Ting Bai
School of Computer Science, Beijing University of Posts and Telecommunications, Beijing, China
Y
Yuan Fang
School of Computing and Information Systems, Singapore Management University, Singapore
L
Lichao Sun
Lehigh University, Bethlehem, Pennsylvania, USA
Philip S. Yu
Philip S. Yu
Professor of Computer Science, University of Illinons at Chicago
Data miningDatabasePrivacy
Chuan Shi
Chuan Shi
Beijing University of Posts and Telecommunications
data miningmachine learningsocial network analysis