HopFormer: Sparse Graph Transformers with Explicit Receptive Field Control

๐Ÿ“… 2026-02-02
๐Ÿ“ˆ Citations: 0
โœจ Influential: 0
๐Ÿ“„ PDF
๐Ÿค– AI Summary
This work proposes HopFormer, a novel graph Transformer that addresses the high computational cost and limited receptive field control of conventional approaches relying on dense global attention and explicit positional encodings. HopFormer introduces, for the first time, a head-specific n-hop masked sparse attention mechanism that explicitly models graph structural information and enables precise control over the receptive fieldโ€”without requiring positional encodings or architectural modifications. The method achieves linear scalability and computational efficiency while delivering competitive or superior performance across diverse node-level and graph-level benchmarks. Notably, it demonstrates that localized attention is more stable and effective than global attention in small-world graphs, challenging the prevailing assumption that graph Transformers inherently depend on global interactions.

Technology Category

Application Category

๐Ÿ“ Abstract
Graph Transformers typically rely on explicit positional or structural encodings and dense global attention to incorporate graph topology. In this work, we show that neither is essential. We introduce HopFormer, a graph Transformer that injects structure exclusively through head-specific n-hop masked sparse attention, without the use of positional encodings or architectural modifications. This design provides explicit and interpretable control over receptive fields while enabling genuinely sparse attention whose computational cost scales linearly with mask sparsity. Through extensive experiments on both node-level and graph-level benchmarks, we demonstrate that our approach achieves competitive or superior performance across diverse graph structures. Our results further reveal that dense global attention is often unnecessary: on graphs with strong small-world properties, localized attention yields more stable and consistently high performance, while on graphs with weaker small-world effects, global attention offers diminishing returns. Together, these findings challenge prevailing assumptions in graph Transformer design and highlight sparsity-controlled attention as a principled and efficient alternative.
Problem

Research questions and friction points this paper is trying to address.

Graph Transformers
sparse attention
receptive field control
positional encodings
graph topology
Innovation

Methods, ideas, or system contributions that make the work stand out.

Sparse Attention
Graph Transformer
Receptive Field Control
n-hop Masking
Positional Encoding-Free
๐Ÿ”Ž Similar Papers
No similar papers found.
Sanggeon Yun
Sanggeon Yun
University of California, Irvine
Machine LearningNeuroSymbolic AIVector Symbolic ArchitectureHyperdimensional Computing
R
Raheeb Hassan
Department of Computer Science, University of California, Irvine
R
Ryozo Masukawa
Department of Computer Science, University of California, Irvine
S
Sungheon Jeong
Department of Computer Science, University of California, Irvine
Mohsen Imani
Mohsen Imani
Associate Professor, University of California Irvine
Machine LearningBrain-Inspired SystemsIntelligent Systems