HeSRN: Representation Learning On Heterogeneous Graphs via Slot-Aware Retentive Network

📅 2025-10-10
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Existing graph Transformers face two key challenges in heterogeneous graph learning: (i) quadratic computational complexity hinders scalability, and (ii) difficulty in disentangling heterogeneous semantics across multiple node types. To address these, we propose the Slot-aware Retention Network (SRN), which introduces a novel slot-aware structural encoder. It decouples type-specific semantics via slot normalization and retention-based fusion—avoiding semantic confusion caused by forced feature-space unification. SRN replaces self-attention with a linear-complexity retention mechanism, augmented by slot projection, distribution alignment, and multi-scale heterogeneous retention layers to jointly capture local structural patterns and global heterogeneous semantics. Evaluated on four real-world heterogeneous graph benchmarks, SRN achieves state-of-the-art performance on node classification—outperforming mainstream GNNs and graph Transformers in both accuracy and computational efficiency.

Technology Category

Application Category

📝 Abstract
Graph Transformers have recently achieved remarkable progress in graph representation learning by capturing long-range dependencies through self-attention. However, their quadratic computational complexity and inability to effectively model heterogeneous semantics severely limit their scalability and generalization on real-world heterogeneous graphs. To address these issues, we propose HeSRN, a novel Heterogeneous Slot-aware Retentive Network for efficient and expressive heterogeneous graph representation learning. HeSRN introduces a slot-aware structure encoder that explicitly disentangles node-type semantics by projecting heterogeneous features into independent slots and aligning their distributions through slot normalization and retention-based fusion, effectively mitigating the semantic entanglement caused by forced feature-space unification in previous Transformer-based models. Furthermore, we replace the self-attention mechanism with a retention-based encoder, which models structural and contextual dependencies in linear time complexity while maintaining strong expressive power. A heterogeneous retentive encoder is further employed to jointly capture both local structural signals and global heterogeneous semantics through multi-scale retention layers. Extensive experiments on four real-world heterogeneous graph datasets demonstrate that HeSRN consistently outperforms state-of-the-art heterogeneous graph neural networks and Graph Transformer baselines on node classification tasks, achieving superior accuracy with significantly lower computational complexity.
Problem

Research questions and friction points this paper is trying to address.

Addressing quadratic complexity and semantic modeling limitations in graph transformers
Disentangling node-type semantics through slot-aware projection and normalization
Capturing structural dependencies with linear-time retention-based encoders
Innovation

Methods, ideas, or system contributions that make the work stand out.

Slot-aware encoder disentangles node-type semantics
Retention-based encoder replaces quadratic self-attention
Multi-scale retention captures local and global dependencies
🔎 Similar Papers
No similar papers found.
Y
Yifan Lu
Mohamed bin Zayed University of Artificial Intelligence, UAE
Z
Ziyun Zou
Mohamed bin Zayed University of Artificial Intelligence, UAE
B
Belal Alsinglawi
Zayed University, UAE; Swinburne University of Technology, Australia
I
Islam Al-Qudah
Higher Colleges of Technology, UAE
I
Izzat Alsmadi
Texas A&M University-San Antonio, United States
F
Feilong Tang
Mohamed bin Zayed University of Artificial Intelligence, UAE; Monash University, Australia
Pengfei Jiao
Pengfei Jiao
Hangzhou Dianzi University
Probabilistic Graphical ModelsGraph Neural NetworkNetwork EmbeddingTemporal NetworksRecommender System
Shoaib Jameel
Shoaib Jameel
Associate Professor, University of Southampton
Machine LearningInformation RetrievalNatural Language ProcessingProbabilistic topic models