Graph2Video: Leveraging Video Models to Model Dynamic Graph Evolution

πŸ“… 2026-03-09
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
This work proposes a novel approach to dynamic graph learning by drawing inspiration from video modeling. Existing methods struggle to effectively capture fine-grained temporal interactions, long-range dependencies, and link-specific relational dynamics in temporal link prediction. To address these limitations, the authors formulate the temporal neighborhood of a target link as a β€œgraph video” and leverage the spatiotemporal inductive biases inherent in foundational video models. This enables unified modeling of both local interaction details and long-term evolutionary patterns, yielding lightweight, plug-and-play link-level embeddings. Extensive experiments on multiple benchmark datasets demonstrate that the proposed method significantly outperforms current state-of-the-art approaches, thereby validating the effectiveness of cross-domain transfer of spatiotemporal modeling capabilities to dynamic graph representation learning.

Technology Category

Application Category

πŸ“ Abstract
Dynamic graphs are common in real-world systems such as social media, recommender systems, and traffic networks. Existing dynamic graph models for link prediction often fall short in capturing the complexity of temporal evolution. They tend to overlook fine-grained variations in temporal interaction order, struggle with dependencies that span long time horizons, and offer limited capability to model pair-specific relational dynamics. To address these challenges, we propose \textbf{Graph2Video}, a video-inspired framework that views the temporal neighborhood of a target link as a sequence of "graph frames". By stacking temporally ordered subgraph frames into a "graph video", Graph2Video leverages the inductive biases of video foundation models to capture both fine-grained local variations and long-range temporal dynamics. It generates a link-level embedding that serves as a lightweight and plug-and-play link-centric memory unit. This embedding integrates seamlessly into existing dynamic graph encoders, effectively addressing the limitations of prior approaches. Extensive experiments on benchmark datasets show that Graph2Video outperforms state-of-the-art baselines on the link prediction task in most cases. The results highlight the potential of borrowing spatio-temporal modeling techniques from computer vision as a promising and effective approach for advancing dynamic graph learning.
Problem

Research questions and friction points this paper is trying to address.

dynamic graphs
link prediction
temporal evolution
graph representation learning
spatio-temporal modeling
Innovation

Methods, ideas, or system contributions that make the work stand out.

dynamic graph
video foundation models
link prediction
temporal modeling
graph representation learning
πŸ”Ž Similar Papers
No similar papers found.