๐ค AI Summary
This work systematically characterizes the expressive power boundaries of Graph Transformers (GTs) and GPS networks under soft and average-hard attention mechanisms. Addressing both idealized (real-number) and practical (floating-point) numerical settings, we establish the first precise logical correspondences: GPS networks are expressively equivalent to Graded Modal Logic (GML) with global modalities over reals, and to its counting extension (CGML) with counting global modalities over floats; GTs correspond instead to propositional logic extended with global modalities in both settings. Leveraging tools from first-order, propositional, and modal logic, we rigorously expose how numerical precision fundamentally constrains logical expressivity. Our framework provides the first fine-grained, formally verifiable logical characterization of graph neural networks, enabling principled analysis of their representational capabilities.
๐ Abstract
Transformers are the basis of modern large language models, but relatively little is known about their precise expressive power on graphs. We study the expressive power of graph transformers (GTs) by Dwivedi and Bresson (2020) and GPS-networks by Rampรกsek et al. (2022), both under soft-attention and average hard-attention. Our study covers two scenarios: the theoretical setting with real numbers and the more practical case with floats. With reals, we show that in restriction to vertex properties definable in first-order logic (FO), GPS-networks have the same expressive power as graded modal logic (GML) with the global modality. With floats, GPS-networks turn out to be equally expressive as GML with the counting global modality. The latter result is absolute, not restricting to properties definable in a background logic. We also obtain similar characterizations for GTs in terms of propositional logic with the global modality (for reals) and the counting global modality (for floats).