Invariant Random Forest: Tree-Based Model Solution for OOD Generalization (2312.04273v3)
Abstract: Out-Of-Distribution (OOD) generalization is an essential topic in machine learning. However, recent research is only focusing on the corresponding methods for neural networks. This paper introduces a novel and effective solution for OOD generalization of decision tree models, named Invariant Decision Tree (IDT). IDT enforces a penalty term with regard to the unstable/varying behavior of a split across different environments during the growth of the tree. Its ensemble version, the Invariant Random Forest (IRF), is constructed. Our proposed method is motivated by a theoretical result under mild conditions, and validated by numerical tests with both synthetic and real datasets. The superior performance compared to non-OOD tree models implies that considering OOD generalization for tree models is absolutely necessary and should be given more attention.
- Invariant risk minimization. arXiv preprint arXiv:1907.02893.
- Breiman, L. 2001. Random forests. Machine learning, 45: 5–32.
- Xgboost: extreme gradient boosting. R package version 0.4-2, 1(4): 1–4.
- Stable learning establishes some common ground between causal inference and machine learning. Nature Machine Intelligence, 4(2): 110–115.
- Friedman, J. H. 2001. Greedy function approximation: a gradient boosting machine. Annals of statistics, 1189–1232.
- Shortcut learning in deep neural networks. Nature Machine Intelligence, 2(11): 665–673.
- Classification and Regression Trees. Biometrics, 40(3): 874.
- Empirical asset pricing via machine learning. The Review of Financial Studies, 33(5): 2223–2273.
- Autoencoder asset pricing models. Journal of Econometrics, 222(1): 429–450.
- Out-of-distribution generalization via risk extrapolation (rex). In International Conference on Machine Learning, 5815–5826. PMLR.
- Stable prediction across unknown environments. In proceedings of the 24th ACM SIGKDD international conference on knowledge discovery & data mining, 1617–1626.
- Stable prediction with model misspecification and agnostic distribution shift. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 34, 4485–4492.
- Decorr: Environment Partitioning for Invariant Learning and OOD Generalization. arXiv preprint arXiv:2211.10054.
- Stable adversarial learning under distributional shifts. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 35, 8662–8670.
- Time Robust Trees: Using Temporal Invariance to Improve Generalization. In Brazilian Conference on Intelligent Systems, 385–397. Springer.
- Quinlan, J. R. 2014. C4. 5: programs for machine learning. Elsevier.
- The risks of invariant risk minimization. arXiv preprint arXiv:2010.05761.
- Stable learning via differentiated variable decorrelation. In Proceedings of the 26th acm sigkdd international conference on knowledge discovery & data mining, 2185–2193.
- Stable learning via sample reweighting. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 34, 5692–5699.
- Towards out-of-distribution generalization: A survey. arXiv preprint arXiv:2108.13624.
- Quantitatively Measuring and Contrastively Exploring Heterogeneity for Domain Generalization. arXiv preprint arXiv:2305.15889.
- Coping with Change: Learning Invariant and Minimum Sufficient Representations for Fine-Grained Visual Categorization. arXiv preprint arXiv:2306.04893.
- Stable Learning via Sparse Variable Independence. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 37, 10998–11006.
- Deep stable learning for out-of-distribution generalization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 5372–5382.
- Domain Generalization: A Survey.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.