LION: A Clifford Neural Paradigm for Multimodal-Attributed Graph Learning

πŸ“… 2026-01-29
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
Existing graph neural approaches struggle to effectively align modalities and fuse information in multimodal attributed graphs, often neglecting graph context, suppressing cross-modal interactions, and lacking adaptive utilization of topological structure. This work introduces Clifford algebra into multimodal graph learning for the first time and proposes a decoupled propagation-aggregation paradigm: it achieves context-aware high-order modality alignment through modality-aware geometric manifolds and designs an adaptive holographic aggregation mechanism based on geometric-grade energy and scale for effective fusion. The proposed method significantly outperforms state-of-the-art models across nine datasets, achieving leading performance on three types of graph tasks and three types of modality tasks.

Technology Category

Application Category

πŸ“ Abstract
Recently, the rapid advancement of multimodal domains has driven a data-centric paradigm shift in graph ML, transitioning from text-attributed to multimodal-attributed graphs. This advancement significantly enhances data representation and expands the scope of graph downstream tasks, such as modality-oriented tasks, thereby improving the practical utility of graph ML. Despite its promise, limitations exist in the current neural paradigms: (1) Neglect Context in Modality Alignment: Most existing methods adopt topology-constrained or modality-specific operators as tokenizers. These aligners inevitably neglect graph context and inhibit modality interaction, resulting in suboptimal alignment. (2) Lack of Adaptation in Modality Fusion: Most existing methods are simple adaptations for 2-modality graphs and fail to adequately exploit aligned tokens equipped with topology priors during fusion, leading to poor generalizability and performance degradation. To address the above issues, we propose LION (c\underline{LI}ff\underline{O}rd \underline{N}eural paradigm) based on the Clifford algebra and decoupled graph neural paradigm (i.e., propagation-then-aggregation) to implement alignment-then-fusion in multimodal-attributed graphs. Specifically, we first construct a modality-aware geometric manifold grounded in Clifford algebra. This geometric-induced high-order graph propagation efficiently achieves modality interaction, facilitating modality alignment. Then, based on the geometric grade properties of aligned tokens, we propose adaptive holographic aggregation. This module integrates the energy and scale of geometric grades with learnable parameters to improve modality fusion. Extensive experiments on 9 datasets demonstrate that LION significantly outperforms SOTA baselines across 3 graph and 3 modality downstream tasks.
Problem

Research questions and friction points this paper is trying to address.

multimodal-attributed graphs
modality alignment
modality fusion
graph context
topology priors
Innovation

Methods, ideas, or system contributions that make the work stand out.

Clifford algebra
multimodal-attributed graphs
modality alignment
adaptive holographic aggregation
decoupled graph neural paradigm
πŸ”Ž Similar Papers
No similar papers found.