Neural network models rely on learning meaningful latent patterns from data for enhanced performance and generalizability.
Self-supervised representation learning for spatial time series data, common in transportation, faces challenges in preserving fine-grained spatio-temporal similarities in the latent space.
A study introduced two structure-preserving regularizers for contrastive learning of spatial time series, balancing the contrastive learning objective and the need for structure preservation with a dynamic weighting mechanism.
The proposed method showed improved performance across tasks like multivariate time series classification and traffic prediction, with enhanced preservation of similarity structures in the latent space.