A robust assessment for invariant representations (2404.05058v1)
Abstract: The performance of machine learning models can be impacted by changes in data over time. A promising approach to address this challenge is invariant learning, with a particular focus on a method known as invariant risk minimization (IRM). This technique aims to identify a stable data representation that remains effective with out-of-distribution (OOD) data. While numerous studies have developed IRM-based methods adaptive to data augmentation scenarios, there has been limited attention on directly assessing how well these representations preserve their invariant performance under varying conditions. In our paper, we propose a novel method to evaluate invariant performance, specifically tailored for IRM-based methods. We establish a bridge between the conditional expectation of an invariant predictor across different environments through the likelihood ratio. Our proposed criterion offers a robust basis for evaluating invariant performance. We validate our approach with theoretical support and demonstrate its effectiveness through extensive numerical studies.These experiments illustrate how our method can assess the invariant performance of various representation techniques.
- “Invariance principle meets information bottleneck for out-of-distribution generalization” In Advances in Neural Information Processing Systems 34, 2021, pp. 3438–3450
- “Invariant risk minimization” In arXiv preprint arXiv:1907.02893, 2019
- Tiffany Tianhui Cai, Hongseok Namkoong and Steve Yadlowsky “Diagnosing model performance under distribution shift” In arXiv preprint arXiv:2303.02011, 2023
- Emmanuel J Candès, Lihua Lei and Zhimei Ren “Conformalized Survival Analysis” In arXiv preprint arXiv:2103.09763, 2021
- “Invariant rationalization” In International Conference on Machine Learning, 2020, pp. 1448–1458 PMLR
- “Robust covariate shift regression” In Artificial Intelligence and Statistics, 2016, pp. 1270–1279 PMLR
- Elliot Creager, Jörn-Henrik Jacobsen and Richard Zemel “Environment inference for invariant learning” In International Conference on Machine Learning, 2021, pp. 2189–2200 PMLR
- “Unsupervised domain adaptation by backpropagation” In International conference on machine learning, 2015, pp. 1180–1189 PMLR
- “Domain-adversarial training of neural networks” In Journal of machine learning research 17.59, 2016, pp. 1–35
- “Out-of-distribution generalization with maximal invariant predictor” In CoRR, 2020
- “Out-of-distribution generalization via risk extrapolation (rex)” In International Conference on Machine Learning, 2021, pp. 5815–5826 PMLR
- “Invariant information bottleneck for domain generalization” In Proceedings of the AAAI Conference on Artificial Intelligence 36.7, 2022, pp. 7399–7407
- “Deep domain generalization via conditional invariant adversarial networks” In Proceedings of the European conference on computer vision (ECCV), 2018, pp. 624–639
- “Bayesian invariant risk minimization” In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 16021–16030
- Divyat Mahajan, Shruti Tople and Amit Sharma “Domain generalization using causal matching” In International conference on machine learning, 2021, pp. 7313–7324 PMLR
- “Input-dependent estimation of generalization error under covariate shift” In Statistics & Risk Modeling 23.4/2005 De Gruyter, 2005, pp. 249–279
- “Representation Learning via Invariant Causal Mechanisms” In International Conference on Learning Representations, 2020
- “Unified deep supervised domain adaptation and generalization” In Proceedings of the IEEE international conference on computer vision, 2017, pp. 5715–5725
- Jonas Peters, Peter Bühlmann and Nicolai Meinshausen “Causal inference by using invariant prediction: identification and confidence intervals” In Journal of the Royal Statistical Society Series B: Statistical Methodology 78.5 Oxford University Press, 2016, pp. 947–1012
- “Dataset shift in machine learning” Mit Press, 2008
- Sashank Reddi, Barnabas Poczos and Alex Smola “Doubly robust covariate shift correction” In Proceedings of the AAAI Conference on Artificial Intelligence 29.1, 2015
- Elan Rosenfeld, Pradeep Ravikumar and Andrej Risteski “The risks of invariant risk minimization” In arXiv preprint arXiv:2010.05761, 2020
- Hidetoshi Shimodaira “Improving predictive inference under covariate shift by weighting the log-likelihood function” In Journal of statistical planning and inference 90.2 Elsevier, 2000, pp. 227–244
- “Domain adaptation with invariant representation learning: What transformations to learn?” In Advances in Neural Information Processing Systems 34, 2021, pp. 24791–24803
- Masashi Sugiyama, Matthias Krauledat and Klaus-Robert Müller “Covariate shift adaptation by importance weighted cross validation.” In Journal of Machine Learning Research 8.5, 2007
- “Input-dependent estimation of generalization error under covariate shift” Oldenbourg Wissenschaftsverlag GmbH, 2005
- “Destination and route attributes associated with adults’ walking: a review.” In Medicine and science in sports and exercise 44.7, 2012, pp. 1275–1286
- “Conformal prediction under covariate shift” In Advances in neural information processing systems 32, 2019
- “Provable domain generalization via invariant-feature subspace recovery” In International Conference on Machine Learning, 2022, pp. 23018–23033 PMLR
- “Federated domain generalization with generalization adjustment” In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023, pp. 3954–3963
- “Nico++: Towards better benchmarking for domain generalization” In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023, pp. 16036–16047
- “Sparse invariant risk minimization” In International Conference on Machine Learning, 2022, pp. 27222–27244 PMLR