OpenMAG: A Comprehensive Benchmark for Multimodal-Attributed Graph

📅 2026-02-05
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the lack of a unified and comprehensive evaluation benchmark for multimodal attribute graph (MAG) models, which has led to significant limitations in domain coverage, encoder flexibility, model diversity, and task scope. To bridge this gap, we propose a standardized benchmark framework encompassing 19 datasets across six domains, 16 encoders, 24 state-of-the-art models, and eight downstream tasks. Our framework supports both static and trainable feature encoders and integrates a curated model library, multimodal encoders, graph neural networks, and a unified training and evaluation pipeline. It enables, for the first time, systematic integration of cross-domain multimodal graph data and diverse architectures, achieving advances in encoder flexibility, task breadth, and evaluation dimensions. Through systematic assessment across five criteria—necessity, data quality, effectiveness, robustness, and efficiency—we derive 14 key insights, offering the community a reproducible, reliable benchmark and empirical guidance.

Technology Category

Application Category

📝 Abstract
Multimodal-Attributed Graph (MAG) learning has achieved remarkable success in modeling complex real-world systems by integrating graph topology with rich attributes from multiple modalities. With the rapid proliferation of novel MAG models capable of handling intricate cross-modal semantics and structural dependencies, establishing a rigorous and unified evaluation standard has become imperative. Although existing benchmarks have facilitated initial progress, they exhibit critical limitations in domain coverage, encoder flexibility, model diversity, and task scope, presenting significant challenges to fair evaluation. To bridge this gap, we present OpenMAG, a comprehensive benchmark that integrates 19 datasets across 6 domains and incorporates 16 encoders to support both static and trainable feature encoding. OpenMAG further implements a standardized library of 24 state-of-the-art models and supports 8 downstream tasks, enabling fair comparisons within a unified framework. Through systematic assessment of necessity, data quality, effectiveness, robustness, and efficiency, we derive 14 fundamental insights into MAG learning to guide future advancements. Our code is available at https://github.com/YUKI-N810/OpenMAG.
Problem

Research questions and friction points this paper is trying to address.

Multimodal-Attributed Graph
benchmark
evaluation standard
model comparison
graph learning
Innovation

Methods, ideas, or system contributions that make the work stand out.

Multimodal-Attributed Graph
comprehensive benchmark
unified evaluation framework
cross-modal semantics
graph representation learning
🔎 Similar Papers
No similar papers found.
C
Chenxi Wan
Department of Computer Science, Beijing Institute of Technology, Beijing, China
Xunkai Li
Xunkai Li
School of Computer Science and Technology, Beijing Institution of Technology
Data-centric AIGraph MLAI4Science
Y
Yilong Zuo
Department of Computer Science, Beijing Institute of Technology, Beijing, China
H
Haokun Deng
Department of Computer Science, Beijing Institute of Technology, Beijing, China
S
Sihan Li
Department of Computer Science, Beijing Institute of Technology, Beijing, China
Bowen Fan
Bowen Fan
UZH
Machine Learning for Healthcare
Hongchao Qin
Hongchao Qin
Beijing Institute of Technology
Graph Data Mining
R
Ronghua Li
Department of Computer Science, Beijing Institute of Technology, Beijing, China
Guoren Wang
Guoren Wang
Beijing Institute of Technology