Hypergraph Foundation Model

📅 2025-03-03
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the dual challenges of jointly modeling higher-order structural dependencies and vertex features in hypergraph data, and the absence of general-purpose foundational models for hypergraphs. To this end, we propose Hyper-FM—the first hypergraph foundation model designed for cross-domain knowledge extraction. Methodologically, we introduce a hierarchical higher-order neighborhood-guided vertex knowledge embedding mechanism and a multi-hypergraph collaborative structural knowledge extraction framework. Moreover, we establish the first scaling law for hypergraph foundation models, revealing that domain diversity contributes more significantly to performance gains than mere parameter or data scale expansion. Evaluated on a newly constructed benchmark of 10 text-attributed hypergraph datasets, Hyper-FM achieves an average 13.3% improvement over state-of-the-art baselines, empirically validating its cross-domain structural generalization capability and effective knowledge transfer—thereby establishing a new paradigm for hypergraph foundation model research.

Technology Category

Application Category

📝 Abstract
Hypergraph neural networks (HGNNs) effectively model complex high-order relationships in domains like protein interactions and social networks by connecting multiple vertices through hyperedges, enhancing modeling capabilities, and reducing information loss. Developing foundation models for hypergraphs is challenging due to their distinct data, which includes both vertex features and intricate structural information. We present Hyper-FM, a Hypergraph Foundation Model for multi-domain knowledge extraction, featuring Hierarchical High-Order Neighbor Guided Vertex Knowledge Embedding for vertex feature representation and Hierarchical Multi-Hypergraph Guided Structural Knowledge Extraction for structural information. Additionally, we curate 10 text-attributed hypergraph datasets to advance research between HGNNs and LLMs. Experiments on these datasets show that Hyper-FM outperforms baseline methods by approximately 13.3%, validating our approach. Furthermore, we propose the first scaling law for hypergraph foundation models, demonstrating that increasing domain diversity significantly enhances performance, unlike merely augmenting vertex and hyperedge counts. This underscores the critical role of domain diversity in scaling hypergraph models.
Problem

Research questions and friction points this paper is trying to address.

Modeling complex high-order relationships in hypergraphs.
Extracting multi-domain knowledge from hypergraph data.
Improving performance through domain diversity in hypergraph models.
Innovation

Methods, ideas, or system contributions that make the work stand out.

Hypergraph Foundation Model for multi-domain knowledge extraction
Hierarchical High-Order Neighbor Guided Vertex Knowledge Embedding
Hierarchical Multi-Hypergraph Guided Structural Knowledge Extraction
🔎 Similar Papers
No similar papers found.
Yifan Feng
Yifan Feng
Assistant Professor, NUS Business School
learninginformationpreferenceplatform and market
S
Shiquan Liu
Institute of Artificial Intelligence and Robotics, College of Artificial Intelligence, Xi’an Jiaotong University, Xi’an 710049, China
Xiangmin Han
Xiangmin Han
Postdoctoral, Tsinghua University
hypergraphmedical image analysisbrain network analysispathology analysis
Shaoyi Du
Shaoyi Du
Institute of Artificial Intelligence and Robotics, Xi'an Jiaotong University
Pattern RecognitionComputer VisionImage Processing
Z
Zongze Wu
College of Mechatronics and Control Engineering, Shenzhen University, Shenzhen, Guangdong 510006, China
H
Han Hu
Beijing Institute of Technology, Beijing 100811, China
Y
Yue Gao
School of Software, BNRist, THUIBCS, BLBCI, Tsinghua University, Beijing 100084, China