MoveGCL is a privacy-preserving framework for training mobility foundation models via generative continual learning without sharing raw data.
It enables decentralized and progressive model evolution by replaying synthetic trajectories generated from a frozen teacher model, while reinforcing knowledge retention through a tailored distillation strategy.
MoveGCL incorporates a Mixture-of-Experts Transformer with a mobility-aware expert routing mechanism and utilizes a layer-wise progressive adaptation strategy to stabilize continual updates.
Experiments on six real-world urban datasets show that MoveGCL achieves performance comparable to joint training, surpassing federated learning baselines, and providing strong privacy protection.