Automatic Domain Adaptation by Transformers in In-Context Learning (2405.16819v1)
Abstract: Selecting or designing an appropriate domain adaptation algorithm for a given problem remains challenging. This paper presents a Transformer model that can provably approximate and opt for domain adaptation methods for a given dataset in the in-context learning framework, where a foundation model performs new tasks without updating its parameters at test time. Specifically, we prove that Transformers can approximate instance-based and feature-based unsupervised domain adaptation algorithms and automatically select an algorithm suited for a given dataset. Numerical results indicate that in-context learning demonstrates an adaptive domain adaptation surpassing existing methods.
- Transformers learn to implement preconditioned gradient descent for in-context learning. Advances in Neural Information Processing Systems, 36, 2024.
- What learning algorithm is in-context learning? investigations with linear models. arXiv preprint arXiv:2211.15661, 2022.
- Transformers as statisticians: Provable in-context learning with in-context algorithm selection. In Thirty-seventh Conference on Neural Information Processing Systems, 2023. URL https://openreview.net/forum?id=liMSqUuVg9.
- A theory of learning from different domains. Machine learning, 79:151–175, 2010.
- Boosting for transfer learning. In Proceedings of the 24th international conference on Machine learning, pp. 193–200, 2007.
- Daumé III, H. Frustratingly easy domain adaptation. arXiv preprint arXiv:0907.1815, 2009.
- Domain-adversarial training of neural networks. Journal of machine learning research, 17(59):1–35, 2016.
- What can transformers learn in-context? a case study of simple function classes. Advances in Neural Information Processing Systems, 35:30583–30598, 2022.
- Domain adaptation for medical image analysis: a survey. IEEE Transactions on Biomedical Engineering, 69(3):1173–1185, 2021.
- Jiang, H. Uniform convergence rates for kernel density estimation. In International Conference on Machine Learning, pp. 1694–1703. PMLR, 2017.
- A least-squares approach to direct importance estimation. The Journal of Machine Learning Research, 10:1391–1445, 2009.
- A short survey on importance weighting for machine learning. Transactions on Machine Learning Research, 2024. ISSN 2835-8856. URL https://openreview.net/forum?id=IhXM3g2gxg. Survey Certification.
- Adam: a Method for Stochastic Optimization. In ICLR, 2015.
- Transformers as algorithms: Generalization and implicit model selection in in-context learning. arXiv preprint arXiv:2301.07067, 2023.
- Transformers as decision makers: Provable in-context reinforcement learning via supervised pretraining. In NeurIPS 2023 Foundation Models for Decision Making Workshop, 2023.
- Decoupled weight decay regularization. In ICLR, 2019. URL https://openreview.net/forum?id=Bkg6RiCqY7.
- Pytorch: An imperative style, high-performance deep learning library. In NeurIPS, 2019.
- Pretraining task diversity and the emergence of non-bayesian in-context learning for regression. Advances in Neural Information Processing Systems, 36, 2024.
- Fast and flexible multi-task classification using conditional neural adaptive processes. Advances in Neural Information Processing Systems, 32, 2019.
- Wasserstein distance guided representation learning for domain adaptation. In Proceedings of the AAAI conference on artificial intelligence, 2018.
- Shimodaira, H. Improving predictive inference under covariate shift by weighting the log-likelihood function. Journal of statistical planning and inference, 90(2):227–244, 2000.
- Covariate shift adaptation by importance weighted cross validation. Journal of Machine Learning Research, 8(5), 2007.
- Density ratio estimation in machine learning. Cambridge University Press, 2012.
- Meta-transfer learning for few-shot learning. In IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2019, Long Beach, CA, USA, June 16-20, 2019, pp. 403–412. Computer Vision Foundation / IEEE, 2019.
- Transformers learn in-context by gradient descent. In International Conference on Machine Learning, pp. 35151–35174. PMLR, 2023.
- An explanation of in-context learning as implicit bayesian inference. In International Conference on Learning Representations, 2021.
- Transfer learning via learning to transfer. In International conference on machine learning, pp. 5085–5094. PMLR, 2018.
- Trained transformers learn linear models in-context. In R0-FoMo: Robustness of Few-shot and Zero-shot Learning in Large Foundation Models, 2023.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.