Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

T-Rep: Representation Learning for Time Series using Time-Embeddings (2310.04486v3)

Published 6 Oct 2023 in cs.LG and cs.AI

Abstract: Multivariate time series present challenges to standard machine learning techniques, as they are often unlabeled, high dimensional, noisy, and contain missing data. To address this, we propose T-Rep, a self-supervised method to learn time series representations at a timestep granularity. T-Rep learns vector embeddings of time alongside its feature extractor, to extract temporal features such as trend, periodicity, or distribution shifts from the signal. These time-embeddings are leveraged in pretext tasks, to incorporate smooth and fine-grained temporal dependencies in the representations, as well as reinforce robustness to missing data. We evaluate T-Rep on downstream classification, forecasting, and anomaly detection tasks. It is compared to existing self-supervised algorithms for time series, which it outperforms in all three tasks. We test T-Rep in missing data regimes, where it proves more resilient than its counterparts. Finally, we provide latent space visualisation experiments, highlighting the interpretability of the learned representations.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (49)
  1. Autowarp: learning a warping distance from unlabeled time series using sequence autoencoders. arXiv preprint arXiv:1810.10107, 2018.
  2. An empirical evaluation of generic convolutional and recurrent networks for sequence modeling. arXiv preprint arXiv:1803.01271, 2018.
  3. Uncovering the structure of clinical eeg signals with self-supervised learning. Journal of Neural Engineering, 18(4):046020, 2021.
  4. A simple framework for contrastive learning of visual representations. In International conference on machine learning, pp.  1597–1607. PMLR, 2020.
  5. Subject-aware contrastive learning for biosignals. arXiv preprint arXiv:2007.04871, 2020.
  6. The ucr time series archive. IEEE/CAA Journal of Automatica Sinica, 6(6):1293–1305, 2019.
  7. Time series change point detection with self-supervised contrastive predictive coding. In Proceedings of the Web Conference 2021, pp.  3124–3135, 2021.
  8. MiniRocket: A very fast (almost) deterministic transform for time series classification. In Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, pp.  248–257, New York, 2021. ACM.
  9. Distribution of the estimators for autoregressive time series with a unit root. Journal of the American statistical association, 74(366a):427–431, 1979.
  10. Time-series representation learning via temporal and contextual contrasting. arXiv preprint arXiv:2106.14112, 2021.
  11. Som-vae: Interpretable discrete representation learning on time series. arXiv preprint arXiv:1806.02199, 2018.
  12. Unsupervised scalable representation learning for multivariate time series. Advances in neural information processing systems, 32, 2019.
  13. PhysioBank, PhysioToolkit, and PhysioNet: components of a new research resource for complex physiologic signals. Circulation, 101(23):E215–20, June 2000.
  14. Masked reconstruction based self-supervision for human activity recognition. In Proceedings of the 2020 ACM International Symposium on Wearable Computers, pp.  45–49, 2020.
  15. Time2vec: Learning a vector representation of time. arXiv preprint arXiv:1907.05321, 2019.
  16. Finding unusual medical time-series subsequences: Algorithms and applications. IEEE Transactions on Information Technology in Biomedicine, 10(3):429–439, 2006.
  17. Towards a rigorous evaluation of time-series anomaly detection. Proceedings of the AAAI Conference on Artificial Intelligence, 36:7194–7201, 06 2022. doi: 10.1609/aaai.v36i7.20680.
  18. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980, 2014.
  19. Auto-encoding variational bayes. arXiv preprint arXiv:1312.6114, 2013.
  20. Clocs: Contrastive learning of cardiac signals across space, time, and patients. In International Conference on Machine Learning, pp.  5606–5615. PMLR, 2021.
  21. Multivariate time series anomaly detection and interpretation using hierarchical inter-metric and temporal embedding. In Proceedings of the 27th ACM SIGKDD conference on knowledge discovery & data mining, pp.  3220–3230, 2021.
  22. Jianhua Lin. Divergence measures based on the shannon entropy. IEEE Transactions on Information theory, 37(1):145–151, 1991.
  23. Timenet: Pre-trained deep recurrent neural network for time series classification. arXiv preprint arXiv:1706.08838, 2017.
  24. UMAP: Uniform Manifold Approximation and Projection for Dimension Reduction. ArXiv e-prints, February 2018.
  25. Meinard Müller. Dynamic time warping. Information retrieval for music and motion, pp.  69–84, 2007.
  26. Saeed Amizadeh Nikolay Laptev, Y. B. A benchmark dataset for time series anomaly detection, 2015. URL https://yahooresearch.tumblr.com/post/114590420346/a-benchmark-dataset-for-time-series-anomaly.
  27. Representation learning with contrastive predictive coding. arXiv preprint arXiv:1807.03748, 2018.
  28. Pytorch: An imperative style, high-performance deep learning library. In Advances in Neural Information Processing Systems 32, pp.  8024–8035. Curran Associates, Inc., 2019.
  29. Scikit-learn: Machine learning in Python. Journal of Machine Learning Research, 12:2825–2830, 2011.
  30. Contrastive learning based self-supervised time-series analysis. Applied Soft Computing, 117:108397, 2022.
  31. Time-series anomaly detection service at microsoft. In Proceedings of the 25th ACM SIGKDD international conference on knowledge discovery & data mining, pp.  3009–3017, 2019.
  32. Early prediction of sepsis from clinical data: The physionet/computing in cardiology challenge 2019. Critical Care Medicine, 48(2), 2020a. ISSN 0090=3493. URL https://journals.lww.com/ccmjournal/fulltext/2020/02000/early_prediction_of_sepsis_from_clinical_data__the.10.aspx.
  33. Early prediction of sepsis from clinical data: the physionet/computing in cardiology challenge 2019. Critical care medicine, 48(2):210–217, 2020b.
  34. Unsupervised pre-training of a deep lstm-based stacked autoencoder for multivariate time series forecasting problems. Scientific reports, 9(1):19038, 2019.
  35. Anomaly detection in streams with extreme value theory. In Proceedings of the 23rd ACM SIGKDD international conference on knowledge discovery and data mining, pp.  1067–1075, 2017.
  36. Robust anomaly detection for multivariate time series through stochastic recurrent neural network. In Proceedings of the 25th ACM SIGKDD international conference on knowledge discovery & data mining, pp.  2828–2837, 2019.
  37. Self-supervised transformer for sparse and irregularly sampled multivariate clinical time-series. ACM Transactions on Knowledge Discovery from Data (TKDD), 16(6):1–17, 2022.
  38. Unsupervised representation learning for time series with temporal neighborhood coding. arXiv preprint arXiv:2106.00750, 2021.
  39. Attention is all you need. Advances in neural information processing systems, 30, 2017.
  40. The curse of dimensionality in data mining and time series prediction. In International work-conference on artificial neural networks, pp.  758–770. Springer, 2005.
  41. Transformers in time series: A survey. arXiv preprint arXiv:2202.07125, 2022.
  42. Random warping series: A random features method for time-series embedding. In International Conference on Artificial Intelligence and Statistics, pp.  793–802. PMLR, 2018.
  43. 10 challenging problems in data mining research. International Journal of Information Technology & Decision Making, 5(04):597–604, 2006.
  44. Ts2vec: Towards universal representation of time series. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 36, pp.  8980–8987, 2022.
  45. A transformer-based framework for multivariate time series representation learning. In Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining, pp.  2114–2124, 2021.
  46. Self-supervised learning for time series analysis: Taxonomy, progress, and prospects. arXiv preprint arXiv:2306.10125, 2023.
  47. Cross reconstruction transformer for self-supervised time series representation learning. arXiv preprint arXiv:2205.09928, 2022.
  48. Lin Yiting Zhihan Yue. Ts2vec github repository, 2021. URL https://github.com/yuezhihan/ts2vec.
  49. Informer: Beyond efficient transformer for long sequence time-series forecasting. In Proceedings of the AAAI conference on artificial intelligence, volume 35, pp.  11106–11115, 2021.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
Citations (8)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com