Transferable Graph Condensation from the Causal Perspective

📅 2026-01-29
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Existing graph dataset compression methods exhibit limited generalization across tasks and domains, often failing to preserve original model performance. This work proposes TGCC, a transferable graph compression framework that introduces causal invariance into graph compression for the first time. TGCC extracts causally invariant features in the spatial domain through causal intervention, integrates enhanced compression operations, and leverages spectral-domain contrastive learning to generate compact datasets that retain the original graph’s causal structure. Extensive experiments on five public benchmarks and a newly constructed FinReport dataset demonstrate that TGCC improves performance by up to 13.41% in cross-task and cross-domain settings and achieves state-of-the-art results in five out of six single-task scenarios.

Technology Category

Application Category

📝 Abstract
The increasing scale of graph datasets has significantly improved the performance of graph representation learning methods, but it has also introduced substantial training challenges. Graph dataset condensation techniques have emerged to compress large datasets into smaller yet information-rich datasets, while maintaining similar test performance. However, these methods strictly require downstream applications to match the original dataset and task, which often fails in cross-task and cross-domain scenarios. To address these challenges, we propose a novel causal-invariance-based and transferable graph dataset condensation method, named \textbf{TGCC}, providing effective and transferable condensed datasets. Specifically, to preserve domain-invariant knowledge, we first extract domain causal-invariant features from the spatial domain of the graph using causal interventions. Then, to fully capture the structural and feature information of the original graph, we perform enhanced condensation operations. Finally, through spectral-domain enhanced contrastive learning, we inject the causal-invariant features into the condensed graph, ensuring that the compressed graph retains the causal information of the original graph. Experimental results on five public datasets and our novel \textbf{FinReport} dataset demonstrate that TGCC achieves up to a 13.41\% improvement in cross-task and cross-domain complex scenarios compared to existing methods, and achieves state-of-the-art performance on 5 out of 6 datasets in the single dataset and task scenario.
Problem

Research questions and friction points this paper is trying to address.

graph condensation
transferability
cross-task
cross-domain
causal invariance
Innovation

Methods, ideas, or system contributions that make the work stand out.

causal invariance
graph condensation
transferable learning
spectral contrastive learning
domain generalization
🔎 Similar Papers
No similar papers found.
H
Huaming Du
Southwestern University of Finance and Economics
Y
Yijie Huang
Southwestern University of Finance and Economics
Su Yao
Su Yao
Tsinghua University
BlockchainCollaborative learning
Y
Yiying Wang
Ant Group
Y
Yueyang Zhou
Ant Group
J
Jingwen Yang
Ant Group
J
Jinshi Zhang
Ant Group
H
Han Ji
Ant Group
Y
Yu Zhao
Southwestern University of Finance and Economics
Guisong Liu
Guisong Liu
university of electronic science and technology of china
neural networks;machine learning;artificial intelligence
H
Hegui Zhang
Dongbei University of Finance and Economics
Carl Yang
Carl Yang
Waymo LLC, PhD at University of California, Davis
GPU ComputingParallel ComputingGraph Processing
Gang Kou
Gang Kou
SWUFE 西南财经大学
Multiple criteria decision makingData miningAHPGroup decision makingOpinion mining