Graph Masked Language Models

📅 2025-02-24
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the challenge of effectively fusing graph-structured and textual information. We propose a dual-branch collaborative learning framework: one branch employs a Graph Neural Network (GNN) to model graph topology, while the other leverages a Pre-trained Language Model (PLM) to capture textual semantics. Our key innovation is a topology-aware semantic masking strategy—dynamically selecting masking targets based on node-theoretic importance—and a learnable soft masking mechanism (feature interpolation) that enables fine-grained alignment between structural and semantic representations. Furthermore, we introduce joint contrastive learning to optimize cross-modal embeddings. Evaluated on multiple node classification and language understanding benchmarks, our method achieves state-of-the-art performance, with significant improvements in model robustness, training convergence stability, and cross-task generalization capability.

Technology Category

Application Category

📝 Abstract
Language Models (LMs) and Graph Neural Networks (GNNs) have shown great promise in their respective areas, yet integrating structured graph data with rich textual information remains challenging. In this work, we propose emph{Graph Masked Language Models} (GMLM), a novel dual-branch architecture that combines the structural learning of GNNs with the contextual power of pretrained language models. Our approach introduces two key innovations: (i) a emph{semantic masking strategy} that leverages graph topology to selectively mask nodes based on their structural importance, and (ii) a emph{soft masking mechanism} that interpolates between original node features and a learnable mask token, ensuring smoother information flow during training. Extensive experiments on multiple node classification and language understanding benchmarks demonstrate that GMLM not only achieves state-of-the-art performance but also exhibits enhanced robustness and stability. This work underscores the benefits of integrating structured and unstructured data representations for improved graph learning.
Problem

Research questions and friction points this paper is trying to address.

Integrating structured graph data with textual information effectively
Combining GNN structural learning with pretrained language models' power
Improving graph learning via structured and unstructured data fusion
Innovation

Methods, ideas, or system contributions that make the work stand out.

Dual-branch architecture combining GNNs and LMs
Semantic masking strategy using graph topology
Soft masking mechanism for smoother training
🔎 Similar Papers
2024-10-09Conference on Empirical Methods in Natural Language ProcessingCitations: 0
Aarush Sinha
Aarush Sinha
University of Copenhagen
Natural Language ProcessingInformation RetrievalMachine LearningMultimodality
C
CU Omkumar
School of Computer Science and Engineering (SCOPE), Vellore Institute of Technology-Chennai, India