TE2Rules: Explaining Tree Ensembles using Rules (2206.14359v5)
Abstract: Tree Ensemble (TE) models, such as Gradient Boosted Trees, often achieve optimal performance on tabular datasets, yet their lack of transparency poses challenges for comprehending their decision logic. This paper introduces TE2Rules (Tree Ensemble to Rules), a novel approach for explaining binary classification tree ensemble models through a list of rules, particularly focusing on explaining the minority class. Many state-of-the-art explainers struggle with minority class explanations, making TE2Rules valuable in such cases. The rules generated by TE2Rules closely approximate the original model, ensuring high fidelity, providing an accurate and interpretable means to understand decision-making. Experimental results demonstrate that TE2Rules scales effectively to tree ensembles with hundreds of trees, achieving higher fidelity within runtimes comparable to baselines. TE2Rules allows for a trade-off between runtime and fidelity, enhancing its practical applicability. The implementation is available here: https://github.com/linkedin/TE2Rules.
- Fast algorithms for mining association rules. In Proc. 20th int. conf. very large data bases, VLDB, volume 1215, 487–499. Citeseer.
- Borgelt, C. 2005. An Implementation of the FP-growth Algorithm. In Proceedings of the 1st international workshop on open source data mining: frequent pattern mining implementations, 1–5.
- Deng, H. 2019. Interpreting tree ensembles with intrees. International Journal of Data Science and Analytics, 7(4): 277–287.
- UCI Machine Learning Repository.
- Predictive learning via rule ensembles. The Annals of Applied Statistics, 2(3): 916 – 954.
- scikit-learn-contrib/skope-rules. https://github.com/scikit-learn-contrib/skope-rules. V1.0.1.
- Why do tree-based models still outperform deep learning on tabular data? arXiv:2207.08815.
- Local Rule-Based Explanations of Black Box Decision Systems. CoRR, abs/1805.10820.
- Making tree ensembles interpretable: A bayesian model selection approach. In International conference on artificial intelligence and statistics, 77–85. PMLR.
- Interpretable decision sets: A joint framework for description and prediction. In Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining, 1675–1684.
- How we analyzed the COMPAS recidivism algorithm. ProPublica.
- Meinshausen, N. 2010. Node harvest. The Annals of Applied Statistics, 4(4).
- Are Neural Rankers still Outperformed by Gradient Boosted Decision Trees? In International Conference on Learning Representations.
- Anchors: High-Precision Model-Agnostic Explanations. In AAAI.
- Born-Again Tree Ensembles. arXiv:2003.11132.
- Falling rule lists. In Artificial intelligence and statistics, 1013–1022. PMLR.
- A bayesian framework for learning rule sets for interpretable classification. The Journal of Machine Learning Research, 18(1): 2357–2393.
- Scalable Bayesian rule lists. In International conference on machine learning, 3921–3930. PMLR.