Breaking Data Silos: Towards Open and Scalable Mobility Foundation Models via Generative Continual Learning

📅 2025-06-07
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address training bottlenecks in human mobility modeling caused by cross-institutional data silos and privacy sensitivity, this paper proposes MoveGCL—the first generative continual learning foundation model tailored for mobility. MoveGCL integrates a frozen teacher-driven trajectory generation module, customized knowledge distillation, and a mobility-aware Mixture-of-Experts (MoE) Transformer routing mechanism, enabling collaborative model evolution without raw data leaving local domains. This design simultaneously mitigates catastrophic forgetting and privacy leakage risks. Extensive experiments across six real-world urban trajectory datasets demonstrate that MoveGCL achieves performance on par with centralized joint training, significantly outperforming state-of-the-art federated learning baselines (average improvement of 12.7%). Moreover, it exhibits strong generalization, scalability, and end-to-end privacy preservation—ensuring compliance with strict data governance requirements while maintaining high predictive fidelity.

Technology Category

Application Category

📝 Abstract
Foundation models have revolutionized fields such as natural language processing and computer vision by enabling general-purpose learning across diverse tasks and datasets. However, building analogous models for human mobility remains challenging due to the privacy-sensitive nature of mobility data and the resulting data silos across institutions. To bridge this gap, we propose MoveGCL, a scalable and privacy-preserving framework for training mobility foundation models via generative continual learning. Without sharing raw data, MoveGCL enables decentralized and progressive model evolution by replaying synthetic trajectories generated from a frozen teacher model, and reinforces knowledge retention through a tailored distillation strategy that mitigates catastrophic forgetting. To address the heterogeneity of mobility patterns, MoveGCL incorporates a Mixture-of-Experts Transformer with a mobility-aware expert routing mechanism, and employs a layer-wise progressive adaptation strategy to stabilize continual updates. Experiments on six real-world urban datasets demonstrate that MoveGCL achieves performance comparable to joint training and significantly outperforms federated learning baselines, while offering strong privacy protection. MoveGCL marks a crucial step toward unlocking foundation models for mobility, offering a practical blueprint for open, scalable, and privacy-preserving model development in the era of foundation models.
Problem

Research questions and friction points this paper is trying to address.

Overcoming data silos in mobility foundation models
Ensuring privacy in decentralized mobility data training
Addressing heterogeneous mobility patterns with adaptive learning
Innovation

Methods, ideas, or system contributions that make the work stand out.

Generative continual learning for mobility models
Privacy-preserving synthetic trajectory replay
Mixture-of-Experts Transformer for mobility patterns
🔎 Similar Papers
No similar papers found.
Y
Yuan Yuan
Tsinghua University
Y
Yukun Liu
Tsinghua University
Chonghua Han
Chonghua Han
Tsinghua University
foundation modelspatio-temporal data mining
J
Jie Feng
Tsinghua University
Y
Yong Li
Tsinghua University