Keywords: self-supervised learning, time series, classification, anomaly detection, forecasting, similarity distillation, teacher-student architecture
Abstract: We propose a self-supervised method for pre-training universal time series representations in which we learn contrastive representations using similarity distillation along the temporal and instance dimensions. We analyze the effectiveness of both dimensions, and evaluate our pre-trained representations on three downstream tasks: time series classification, anomaly detection, and forecasting.
TL;DR: We propose a self-supervised methods for pre-training universal time series representations using similarity distillation.
Community Implementations: [![CatalyzeX](/images/catalyzex_icon.svg) 8 code implementations](https://www.catalyzex.com/paper/self-supervised-time-series-representation/code)
0 Replies
Loading