Prompting-based Temporal Domain Generalization (2310.02473v2)
Abstract: Machine learning traditionally assumes that the training and testing data are distributed independently and identically. However, in many real-world settings, the data distribution can shift over time, leading to poor generalization of trained models in future time periods. This paper presents a novel prompting-based approach to temporal domain generalization that is parameter-efficient, time-efficient, and does not require access to future data during training. Our method adapts a trained model to temporal drift by learning global prompts, domain-specific prompts, and drift-aware prompts that capture underlying temporal dynamics. Experiments on classification, regression, and time series forecasting tasks demonstrate the generality of the proposed approach. The code repository will be publicly shared.
- Self-adaptive forecasting for improved deep learning on non-stationary time-series. arXiv preprint arXiv:2202.02403, 2022.
- Attempt: Parameter-efficient multi-task tuning via attentional mixtures of soft prompts. In Empirical Methods in Natural Language Processing, 2022.
- Temporal domain generalization with drift-aware dynamic neural networks. In International Conference on Learning Representations, 2023.
- A theory of learning from different domains. Machine learning, 2010.
- Language models are few-shot learners. Advances in neural information processing systems, 2020.
- Optimal transport for domain adaptation. IEEE transactions on pattern analysis and machine intelligence, 2016.
- Bert: Pre-training of deep bidirectional transformers for language understanding. North American Chapter of the Association for Computational Linguistics, 2018.
- Adaptive methods for real-world domain generalization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021.
- Adversarially adaptive normalization for single domain generalization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021.
- Unsupervised domain adaptation by backpropagation. In International conference on machine learning, 2015.
- Domain-adversarial training of neural networks. The journal of machine learning research, 2016.
- Making pre-trained language models better few-shot learners. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, 2021.
- Learn to expect the unexpected: Probably approximately correct domain generalization. In International Conference on Artificial Intelligence and Statistics, 2021.
- Geodesic flow kernel for unsupervised domain adaptation. In IEEE conference on computer vision and pattern recognition, 2012.
- Ppt: Pre-trained prompt tuning for few-shot learning. In Proceedings of Annual Meeting of the Association for Computational Linguistic, 2022.
- Bertese: Learning to speak to bert. Conference of the European Chapter of the Association for Computational Linguistics, 2021.
- Continuous manifold based adaptation for evolving visual domains. In Conference on Computer Vision and Pattern Recognition, 2014.
- LoRA: Low-rank adaptation of large language models. In International Conference on Learning Representations, 2022.
- Domain and content adaptive convolution for domain generalization in medical image segmentation. IEEE Transactions on Medical Imaging, 2023.
- How can we know what language models know? Transactions of the Association for Computational Linguistics, 2020.
- Cdot: Continuous domain adaptation using optimal transport. Optimal Transport & Machine Learning Workhop (NeurIPS), 2019.
- Domain randomization for scene-specific car detection and pose estimation. In Winter Conference on Applications of Computer Vision, 2019.
- Self-balanced learning for domain generalization. In International Conference on Image Processing, 2021.
- Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980, 2014.
- Continuous domain adaptation with variational domain-agnostic feature replay. arXiv preprint arXiv:2003.04382, 2020.
- The power of scale for parameter-efficient prompt tuning. CoRR, 2021.
- Bart: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. In Jurafsky, D., Chai, J., Schluter, N., and Tetreault, J. R. (eds.), Proceedings of the nnual Meeting of the Association for Computational Linguistics, 2020.
- Prefix-tuning: Optimizing continuous prompts for generation. Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, 2021.
- A unified feature disentangler for multi-domain image translation and manipulation. Advances in neural information processing systems, 2018.
- Pre-train, prompt, and predict: A systematic survey of prompting methods in natural language processing. ACM Computing Surveys, 2023a.
- Gpt understands, too. AI Open, 2023b.
- Oscillation and chaos in physiological control systems. Science, 1977.
- Adagraph: Unifying predictive and continuous domain adaptation through graphs. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2019.
- Representation learning via invariant causal mechanisms. In International Conference on Learning Representations, 2021.
- Reducing domain gap by reducing style bias. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021.
- Training for the future: A simple gradient interpolation loss to generalize along time. Advances in Neural Information Processing Systems, 2021.
- Domain generalization using shape representation. In European Conference on Computer Vision, 2020.
- Deep contextualized word representations. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics, 2018.
- Adapterfusion: Non-destructive task composition for transfer learning. Annual Meeting of the Association for Computational Linguistics, 2021.
- Structured domain randomization: Bridging the reality gap by context-aware synthetic data. In International Conference on Robotics and Automation, 2019.
- Unsupervised domain generalization for person re-identification: A domain-specific adaptive framework. arXiv preprint arXiv:2111.15077, 2021.
- Learning to learn single domain generalization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020.
- Exploring the limits of transfer learning with a unified text-to-text transformer. Journal of Machine Learning Research, 2020.
- Fishr: Invariant gradient variances for out-of-distribution generalization. International Conference on Machine Learning, 2022.
- How many data points is a prompt worth? In Conference of the North American Chapter of the Association for Computational Linguistics, 2021.
- Exploiting cloze-questions for few-shot text classification and natural language inference. In Proceedings of the Conference of the European Chapter of the Association for Computational Linguistics, 2021.
- Scaleformer: Iterative multi-scale refining transformers for time series forecasting. International Conference on Learning Representations, 2023.
- Generalizing across domains via cross-gradient training. International Conference on Learning Representations, 2018.
- AutoPrompt: Eliciting knowledge from language models with automatically generated prompts. In Empirical Methods in Natural Language Processing, 2020.
- On transferability of prompt tuning for natural language processing. Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 2021.
- Conditioned natural language generation using only unconditioned language model: An exploration. arXiv preprint arXiv:2011.07347, 2020.
- Recovering latent causal factor for generalization to distributional shifts. In Conference on Neural Information Processing Systems, 2021.
- Neuron coverage-guided domain generalization. The IEEE Transactions on Pattern Analysis and Machine Intelligence, 2022.
- Training deep networks with synthetic data: Bridging the reality gap by domain randomization. In Proceedings of the IEEE conference on computer vision and pattern recognition workshops, 2018.
- Learning a universal template for few-shot dataset generalization. In International conference on machine learning, 2021.
- Adversarial discriminative domain adaptation. In Proceedings of the IEEE conference on computer vision and pattern recognition, 2017.
- Generalizing to unseen domains via adversarial data augmentation. Conference on Neural Information Processing Systems, 2018.
- Spot: Better frozen model adaptation through soft prompt transfer. Association for Computational Linguistics, 2021.
- Meta-learning for domain generalization in semantic parsing. In North American Chapter of the Association for Computational Linguistics, 2021a.
- TransPrompt: Towards an automatic transferable prompting framework for few-shot text classification. In Proceedings of the Conference on Empirical Methods in Natural Language Processing, 2021b.
- Continuously indexed domain adaptation. International Conference on Machine Learning, 2020a.
- Generalizing to unseen domains: A survey on domain generalization. arXiv preprint arXiv:2103.03097, 2021c.
- Deep visual domain adaptation: A survey. Neurocomputing, 2018.
- K-adapter: Infusing knowledge into pre-trained models with adapters. Association for Computational Linguistics, 2020b.
- Multitask prompt tuning enables parameter-efficient transfer learning. In International Conference on Learning Representations, 2023.
- Collaborative optimization and aggregation for decentralized domain generalization and adaptation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 2021.
- Multivariate regression on the grassmannian for predicting novel domains. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2016.
- Spatial-temporal change trend analysis of second-hand house price in hefei based on spatial network. Computational Intelligence and Neuroscience, 2022.
- Domain randomization and pyramid consistency: Simulation-to-real generalization without accessing target domain data. In International Conference on Computer Vision, 2019.
- Learning to generalize unseen domains via memory-based multi-source meta-learning for person re-identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021.
- Informer: Beyond efficient transformer for long sequence time-series forecasting. In Proceedings of Association for the Advancement of Artificial Intelligence, 2021.
- Deep domain-adversarial image generation for domain generalisation. In Proceddings of Association for the Advancement of Artificial Intelligence, 2020.