Towards Understanding Variants of Invariant Risk Minimization through the Lens of Calibration (2401.17541v4)
Abstract: Machine learning models traditionally assume that training and test data are independently and identically distributed. However, in real-world applications, the test distribution often differs from training. This problem, known as out-of-distribution (OOD) generalization, challenges conventional models. Invariant Risk Minimization (IRM) emerges as a solution that aims to identify invariant features across different environments to enhance OOD robustness. However, IRM's complexity, particularly its bi-level optimization, has led to the development of various approximate methods. Our study investigates these approximate IRM techniques, using the consistency and variance of calibration across environments as metrics to measure the invariance aimed for by IRM. Calibration, which measures the reliability of model prediction, serves as an indicator of whether models effectively capture environment-invariant features by showing how uniformly over-confident the model remains across varied environments. Through a comparative analysis of datasets with distributional shifts, we observe that Information Bottleneck-based IRM achieves consistent calibration across different environments. This observation suggests that information compression techniques, such as IB, are potentially effective in achieving model invariance. Furthermore, our empirical evidence indicates that models exhibiting consistent calibration across environments are also well-calibrated. This demonstrates that invariance and cross-environment calibration are empirically equivalent. Additionally, we underscore the necessity for a systematic approach to evaluating OOD generalization. This approach should move beyond traditional metrics, such as accuracy and F1 scores, which fail to account for the model's degree of over-confidence, and instead focus on the nuanced interplay between accuracy, calibration, and model invariance.
- Invariant risk minimization games. arXiv preprint arXiv:2002.04692, 2020.
- Invariance principle meets information bottleneck for out-of-distribution generalization. arXiv preprint arXiv:2106.06607, 2022.
- Invariant risk minimization. arXiv preprint arXiv:1907.02893, 2020.
- Pareto invariant risk minimization: Towards mitigating the optimization dilemma in out-of-distribution generalization. arXiv preprint arXiv:2206.07766, 2022.
- Unbiased metric learning: On the utilization of multiple datasets and web images for softening bias. In Proceedings of the IEEE International Conference on Computer Vision, pp. 1657–1664, 2013.
- Domain-adversarial training of neural networks. The journal of machine learning research, 17(1):2096–2030, 2016.
- Shortcut learning in deep neural networks. Nature Machine Intelligence, 2(11):665–673, 2020.
- Domain generalization for object recognition with multi-task autoencoders. In Proceedings of the IEEE international conference on computer vision, pp. 2551–2559, 2015.
- In search of lost domain generalization. arXiv preprint arXiv:2007.01434, 2020.
- On calibration of modern neural networks. In International Conference on Machine Learning, pp. 1321–1330, 2017a.
- On calibration of modern neural networks. In Doina Precup and Yee Whye Teh (eds.), Proceedings of the 34th International Conference on Machine Learning, volume 70 of Proceedings of Machine Learning Research, pp. 1321–1330. PMLR, 06–11 Aug 2017b. URL https://proceedings.mlr.press/v70/guo17a.html.
- Lora: Low-rank adaptation of large language models, 2021.
- The missing invariance principle found – the reciprocal twin of invariant risk minimization, 2023.
- Adam: A method for stochastic optimization. In International Conference on Learning Representations, 2015.
- Unpacking information bottlenecks: Unifying information-theoretic objectives in deep learning, 2021.
- Out-of-distribution generalization via risk extrapolation (rex). arXiv preprint arXiv:2003.00688, 2021.
- Deeper, broader and artier domain generalization. In Proceedings of the IEEE international conference on computer vision, pp. 5542–5550, 2017.
- Bayesian invariant risk minimization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 16021–16030, 2022.
- An empirical investigation of pre-trained model selection for out-of-distribution generalization and calibration. arXiv preprint arXiv:2307.08187, 2023.
- Can you trust your model’s uncertainty? evaluating predictive uncertainty under dataset shift. In Advances in neural information processing systems, volume 32, 2019.
- Dataset shift in machine learning. Mit Press, 2008.
- Learning transferable visual models from natural language supervision. In International conference on machine learning, pp. 8748–8763. PMLR, 2021.
- Optimal representations for covariate shift. arXiv preprint arXiv:2201.00057, 2021.
- Vladimir Vapnik. Principles of risk minimization for learning theory. Advances in neural information processing systems, 4, 1991.
- On calibration and out-of-domain generalization. In Advances in Neural Information Processing Systems, 2021a.
- On calibration and out-of-domain generalization. In A. Beygelzimer, Y. Dauphin, P. Liang, and J. Wortman Vaughan (eds.), Advances in Neural Information Processing Systems, 2021b. URL https://openreview.net/forum?id=XWYJ25-yTRS.
- Nanyang Ye et al. Ood-bench: Quantifying and understanding two dimensions of out-of-distribution generalization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022.
- What is missing in irm training and evaluation? challenges and solutions. arXiv preprint arXiv:2303.02343, 2023a.
- What is missing in irm training and evaluation? challenges and solutions, 2023b.