Unified Transfer Learning Models in High-Dimensional Linear Regression (2307.00238v4)
Abstract: Transfer learning plays a key role in modern data analysis when: (1) the target data are scarce but the source data are sufficient; (2) the distributions of the source and target data are heterogeneous. This paper develops an interpretable unified transfer learning model, termed as UTrans, which can detect both transferable variables and source data. More specifically, we establish the estimation error bounds and prove that our bounds are lower than those with target data only. Besides, we propose a source detection algorithm based on hypothesis testing to exclude the nontransferable data. We evaluate and compare UTrans to the existing algorithms in multiple experiments. It is shown that UTrans attains much lower estimation and prediction errors than the existing methods, while preserving interpretability. We finally apply it to the US intergenerational mobility data and compare our proposed algorithms to the classical machine learning algorithms.
- Fast global convergence of gradient methods for high-dimensional statistical recovery. The Annals of Statistics, 40(5):2452 – 2482, 2012.
- Alexis Bellot and Mihaela van der Schaar. Boosting transfer learning with survival data from heterogeneous domains. In The 22nd International Conference on Artificial Intelligence and Statistics, pages 57–65. PMLR, 2019.
- Simultaneous analysis of lasso and dantzig selector. The Annals of Statistics, 37(4):1705–1732, 2009.
- Coordinate descent algorithms for nonconvex penalized regression, with applications to biological feature selection. The Annals of Applied Statistics, 5(1):232, 2011.
- Transfer learning for nonparametric classification: Minimax rate and adaptive classifier. The Annals of Statistics, 49(1):100–128, 2021.
- Testing generalized linear models with high-dimensional nuisance parameters. Biometrika, 2022.
- Variable selection via nonconcave penalized likelihood and its oracle properties. Journal of the American Statistical Association, 96(456):1348–1360, 2001.
- Testing against a high-dimensional alternative in the generalized linear model: asymptotic type i error control. Biometrika, pages 381–390, 2011.
- Data shared lasso: A novel tool to discover uplift. Computational Statistics & Data Analysis, 101:226–235, 2016.
- Tests for high dimensional generalized linear models. Journal of the Royal Statistical Society: Series B (Statistical Methodology), 78(5):1079–1102, 2016.
- Transfer learning for high-dimensional linear regression: Prediction, estimation and minimax optimality. Journal of the Royal Statistical Society. Series B, Statistical Methodology, 84(1):149—173, 2022.
- On transfer learning in functional linear regression. arXiv preprint arXiv:2206.04277, 2022.
- Adaptive weighted multi-view clustering. In Conference on Health, Inference, and Learning, pages 19–36. PMLR, 2023.
- Multiple-splitting projection test for high-dimensional mean vectors. Journal of Machine Learning Research, 23(71):1–27, 2022.
- Regularized m-estimators with nonconvexity: Statistical and algorithmic theory for local optima. Journal of Machine Learning Research, 16(19):559–616, 2015.
- Handbook of Research on Machine Learning Applications and Trends: Algorithms, Methods, and Techniques. IGI global, 2009.
- Regression modelling on stratified data with the lasso. Biometrika, 104(1):83–96, 2017.
- A survey on transfer learning. IEEE Transactions on Knowledge and Data Engineering, 22(10):1345–1359, 2009.
- R Core Team. R: A Language and Environment for Statistical Computing. R Foundation for Statistical Computing, Vienna, Austria, 2024. URL https://www.R-project.org/.
- Adaptive transfer learning. The Annals of Statistics, 49(6):3618–3649, 2021.
- Ye Tian and Yang Feng. Transfer learning under high-dimensional generalized linear models. Journal of the American Statistical Association, pages 1–14, 2022.
- Robert Tibshirani. Regression shrinkage and selection via the lasso. Journal of the Royal Statistical Society: Series B (Methodological), 58(1):267–288, 1996.
- On the theory of transfer learning: The importance of task diversity. Advances in Neural Information Processing Systems, 33:7852–7862, 2020.
- Sara A Van De Geer and Peter Bühlmann. On the conditions used to prove oracle results for the lasso. Electronic Journal of Statistics, 3:1360–1392, 2009.
- Roman Vershynin. Introduction to the non-asymptotic analysis of random matrices. arXiv preprint arXiv:1011.3027, 2010.
- Transfer learning via learning to transfer. In 35th International Conference on Machine Learning, ICML 2018, volume 11, page 8059, 2018.
- Efficient transfer learning method for automatic hyperparameter tuning. In Artificial Intelligence and Statistics, pages 1077–1085. PMLR, 2014.
- Cun-Hui Zhang. Nearly unbiased variable selection under minimax concave penalty. The Annals of Statistics, 38(2):894–942, 2010.
- A general theory of concave regularization for high-dimensional sparse estimation problems. Statistical Science, 27(4):576–593, 2012.
- A comprehensive survey on transfer learning. Proceedings of the IEEE, 109(1):43–76, 2020.