Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
125 tokens/sec
GPT-4o
47 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Low-Rank Adaptation of Time Series Foundational Models for Out-of-Domain Modality Forecasting (2405.10216v1)

Published 16 May 2024 in cs.LG, cs.AI, and eess.SP

Abstract: Low-Rank Adaptation (LoRA) is a widely used technique for fine-tuning large pre-trained or foundational models across different modalities and tasks. However, its application to time series data, particularly within foundational models, remains underexplored. This paper examines the impact of LoRA on contemporary time series foundational models: Lag-Llama, MOIRAI, and Chronos. We demonstrate LoRA's fine-tuning potential for forecasting the vital signs of sepsis patients in intensive care units (ICUs), emphasizing the models' adaptability to previously unseen, out-of-domain modalities. Integrating LoRA aims to enhance forecasting performance while reducing inefficiencies associated with fine-tuning large models on limited domain-specific data. Our experiments show that LoRA fine-tuning of time series foundational models significantly improves forecasting, achieving results comparable to state-of-the-art models trained from scratch on similar modalities. We conduct comprehensive ablation studies to demonstrate the trade-offs between the number of tunable parameters and forecasting performance and assess the impact of varying LoRA matrix ranks on model performance.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (32)
  1. Chronos: Learning the language of time series. arXiv preprint arXiv:2403.07815 (2024).
  2. Foundational models defining a new era in vision: A survey and outlook. arXiv preprint arXiv:2307.13721 (2023).
  3. Parameter Efficient Fine Tuning: A Comprehensive Analysis Across Applications. arXiv preprint arXiv:2404.13506 (2024).
  4. Strong Baselines for Parameter-Efficient Few-Shot Fine-Tuning. AAAI Conference on Artificial Intelligence (2024).
  5. A transformer architecture for stress detection from ecg. ACM International Symposium on Wearable Computers (2021).
  6. AttX: Attentive cross-connections for fusion of wearable signals in emotion recognition. ACM Transactions on Computing for Healthcare (2024).
  7. Vital sign forecasting for sepsis patients in ICUs. IEEE International Conference on Artificial Intelligence for Medicine, Health and Care (2024).
  8. Interpreting forecasted vital signs using n-beats in sepsis patients. arXiv preprint arXiv:2306.14016 (2023).
  9. On the opportunities and risks of foundation models. arXiv preprint arXiv:2108.07258 (2021).
  10. Your Day in Your Pocket: Complex Activity Recognition from Smartphone Accelerometers. International Conference on Pervasive Computing Technologies for Healthcare (2022).
  11. Prediction of late-onset sepsis in preterm infants using monitoring signals and machine learning. Critical Care Explorations (2021).
  12. Parameter-Efficient Fine-Tuning for Medical Image Analysis: The Missed Opportunity. Medical Imaging with Deep Learning (2024).
  13. Azul Garza and Max Mergenthaler-Canseco. 2023. TimeGPT-1. arXiv preprint arXiv:2310.03589 (2023).
  14. Large language models are zero-shot time series forecasters. Advances in Neural Information Processing Systems (2024).
  15. Divij Gupta and Ali Etemad. 2023. Remote Heart Rate Monitoring in Smart Environments from Videos with Self-supervised Pre-training. IEEE Internet of Things Journal (2023).
  16. Parameter-efficient transfer learning for NLP. International Conference on Machine Learning (2019).
  17. LoRA: Low-Rank Adaptation of Large Language Models. International Conference on Learning Representations (2021).
  18. AdvLoRA: Adversarial Low-Rank Adaptation of Vision-Language Models. arXiv preprint arXiv:2404.13425 (2024).
  19. GPT4MTS: Prompt-based Large Language Model for Multimodal Time-series Forecasting. AAAI Conference on Artificial Intelligence (2024).
  20. Exploring Versatile Generative Language Model Via Parameter-Efficient Transfer Learning. Findings of the Association for Computational Linguistics (2020).
  21. Foundation and large language models: fundamentals, challenges, opportunities, and social impacts. Cluster Computing (2024).
  22. The eICU Collaborative Research Database, a freely available multi-center database for critical care research. Scientific data (2018).
  23. EEG-based Cognitive Load Classification using Feature Masked Autoencoding and Emotion Transfer Learning. ACM International Conference on Multimodal Interaction (2023).
  24. Exploring the limits of transfer learning with a unified text-to-text transformer. Journal of Machine Learning Research (2020).
  25. Lag-llama: Towards foundation models for time series forecasting. arXiv preprint arXiv:2310.08278 (2023).
  26. AdapterDrop: On the Efficiency of Adapters in Transformers. Conference on Empirical Methods in Natural Language Processing (2021).
  27. Three things everyone should know about vision transformers. European Conference on Computer Vision (2022).
  28. Llama: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971 (2023).
  29. Attention is all you need. Advances in Neural Information Processing Systems (2017).
  30. An analysis of physiological and psychological responses in virtual reality and flat screen gaming. IEEE Transactions on Affective Computing (2024).
  31. Unified training of universal time series forecasting transformers. arXiv preprint arXiv:2402.02592 (2024).
  32. BitFit: Simple Parameter-efficient Fine-tuning for Transformer-based Masked Language-models. Annual Meeting of the Association for Computational Linguistics (2022).
Citations (1)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets