Fair Feature Importance Scores for Interpreting Tree-Based Methods and Surrogates (2310.04352v1)
Abstract: Across various sectors such as healthcare, criminal justice, national security, finance, and technology, large-scale ML and AI systems are being deployed to make critical data-driven decisions. Many have asked if we can and should trust these ML systems to be making these decisions. Two critical components are prerequisites for trust in ML systems: interpretability, or the ability to understand why the ML system makes the decisions it does, and fairness, which ensures that ML systems do not exhibit bias against certain individuals or groups. Both interpretability and fairness are important and have separately received abundant attention in the ML literature, but so far, there have been very few methods developed to directly interpret models with regard to their fairness. In this paper, we focus on arguably the most popular type of ML interpretation: feature importance scores. Inspired by the use of decision trees in knowledge distillation, we propose to leverage trees as interpretable surrogates for complex black-box ML models. Specifically, we develop a novel fair feature importance score for trees that can be used to interpret how each feature contributes to fairness or bias in trees, tree-based ensembles, or tree-based surrogates of any complex ML system. Like the popular mean decrease in impurity for trees, our Fair Feature Importance Score is defined based on the mean decrease (or increase) in group bias. Through simulations as well as real examples on benchmark fairness datasets, we demonstrate that our Fair Feature Importance Score offers valid interpretations for both tree-based ensembles and tree-based surrogates of other ML systems.
- “A Reductions Approach to Fair Classification” In ICML 2018: Proceedings of the 35th International Conference on Machine Learning 80, 2018, pp. 60–69 URL: https://proceedings.mlr.press/v80/agarwal18a.html
- Sushant Agarwal “Trade-offs between fairness and interpretability in machine learning” In IJCAI 2021 Workshop on AI for Social Good, 2021
- Genevera I Allen, Luqin Gan and Lili Zheng “Interpretable Machine Learning for Discovery: Statistical Challenges\\\backslash\& Opportunities” In arXiv preprint arXiv:2308.01475, 2023
- “Explainability for fair machine learning” In arXiv preprint arXiv:2010.07389, 2020
- “Data Decisions and Theoretical Implications when Adversarially Learning Fair Representations” In ArXiv Pre-Print 1707.00075, 2017 DOI: 10.48550/arXiv.1707.00075
- Vaishnavi Bhargava, Miguel Couceiro and Amedeo Napoli “LimeOut: an ensemble approach to improve process fairness” In Joint European conference on machine learning and knowledge discovery in databases, 2020, pp. 475–491 Springer
- “Machine learning explainability through comprehensible decision trees” In Machine Learning and Knowledge Extraction: Third IFIP TC 5, TC 12, WG 8.4, WG 8.9, WG 12.9 International Cross-Domain Conference, CD-MAKE 2019, Canterbury, UK, August 26–29, 2019, Proceedings 3, 2019, pp. 15–26 Springer
- Leo Breiman “Bagging Predictors” In Machine Learning 24.2 Springer, 1996, pp. 123–140 DOI: 10.1023/A:1018054314350
- Leo Breiman “Classification and regression trees” Routledge, 1973
- “Toward a taxonomy of trust for probabilistic machine learning” In Science Advances 9.7 American Association for the Advancement of Science, 2023, pp. eabn3999
- Cristian Buciluǎ, Rich Caruana and Alexandru Niculescu-Mizil “Model compression” In Proceedings of the 12th ACM SIGKDD international conference on Knowledge discovery and data mining, 2006, pp. 535–541
- “Optimized pre-processing for discrimination prevention” In NeurIPS 2017: Advances in Neural Information Processing Systems 30, 2017, pp. 3995–4004 DOI: https://proceedings.neurips.cc/paper/2017/hash/9a49a25d845a483fae4be7e341368e36-Abstract.html
- Ian Carlos Campbell “The Apple Card Doesn’t Actually Discriminate Against Women, Investigators Say” In The Verge, 2021 URL: https://www.theverge.com/2021/3/23/22347127/goldman-sachs-apple-card-no-gender-discrimination
- “Fairness in machine learning: A survey” In ArXiv Pre-Print 2010.04053, 2020 DOI: 10.48550/arXiv.2010.04053
- “L-shapley and c-shapley: Efficient model interpretation for structured data” In arXiv preprint arXiv:1808.02610, 2018
- “The Frontiers of Fairness in Machine Learning” In ArXiv Pre-Print 1810.08810, 2018 DOI: 10.48550/arXiv.1810.08810
- Mengnan Du, Ninghao Liu and Xia Hu “Techniques for interpretable machine learning” In Communications of the ACM 63.1 ACM New York, NY, USA, 2019, pp. 68–77
- “UCI Machine Learning Repository”, 2017 URL: http://archive.ics.uci.edu/ml
- “A Comparative Study of Fairness-Enhancing Interventions in Machine Learning” In FAccT 2019: Proceedings of the 2019 Conference on Fairness, Accountability, and Transparency, 2019, pp. 329–338 DOI: 10.1145/3287560.3287589
- “Knowledge distillation: A survey” In International Journal of Computer Vision 129 Springer, 2021, pp. 1789–1819
- “Fair Adversarial Gradient Tree Boosting” In ICDM 2019: Proceedings of the 2019 IEEE International Conference on Data Mining, 2019, pp. 1060–1065 DOI: 10.1109/ICDM.2019.00124
- “A Survey of Methods for Explaining Black Box Models” In ACM Computing Surveys 51.5 New York, NY, USA: Association for Computing Machinery, 2018 DOI: 10.1145/3236009
- Moritz Hardt, Eric Price and Nathan Srebron “Equality of Opportunity in Supervised Learning” In NeurIPS 2016: Advances in Neural Information Processing Systems 29 29, 2016 URL: https://papers.nips.cc/paper/2016/hash/9d2682367c3935defcb1f9e247a97c0d-Abstract.html
- Geoffrey Hinton, Oriol Vinyals and Jeff Dean “Distilling the knowledge in a neural network” In arXiv preprint arXiv:1503.02531, 2015
- Will Knight “The Apple Card Didn’t See Gender—and That’s the Problem” In WIRED, 2019 URL: https://www.wired.com/story/the-apple-card-didnt-see-genderand-thats-the-problem/
- Jeff Larson, Marjorie Roswell and Vaggelis Atlidakis “COMPAS Recidivism Risk Score Data and Analysis”, 2022 URL: https://github.com/propublica/compas-analysis/
- Zachary C Lipton “The mythos of model interpretability: In machine learning, the concept of interpretability is both important and slippery.” In Queue 16.3, 2018, pp. 31–57
- Camille Olivia Little, Michael Weylandt and Genevera I. Allen “To the Fairness Frontier and Beyond: Identifying, Quantifying, and Optimizing the Fairness-Accuracy Pareto Frontier” In arXiv preprint arXiv:2206.00074, 2022
- Wei-Yin Loh “Classification and regression trees” In Wiley interdisciplinary reviews: data mining and knowledge discovery Wiley Online Library, 2011, pp. 14–23
- “Bias Mitigation Post-processing for Individual and Group Fairness” In ICASSP 2019: Proceedings of the 2019 IEEE International Conference on Acoustics, Speech and Signal Processing, 2019, pp. 2847–2851 DOI: 10.1109/ICASSP.2019.8682620
- “Understanding variable importances in forests of randomized trees” In Advances in neural information processing systems 26, 2013
- Masayoshi Mase, Art B Owen and Benjamin B Seiler “Cohort Shapley value for algorithmic fairness” In arXiv preprint arXiv:2105.07168, 2021
- Christoph Molnar “Interpretable machine learning” 2 edn., 2020
- “Definitions, methods, and applications in interpretable machine learning” In Proceedings of the National Academy of Sciences 116.44, 2019, pp. 22071–22080
- “Explainable, trustworthy, and ethical machine learning for healthcare: A survey” In Computers in Biology and Medicine Elsevier, 2022, pp. 106043
- Marco Tulio Ribeiro, Sameer Singh and Carlos Guestrin “"Why Should I Trust You?": Explaining the Predictions of Any Classifier” In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining Association for Computing Machinery, 2016, pp. 1135–1144 DOI: 10.1145/2939672.2939778
- Cynthia Rudin, Caroline Wang and Beau Coker “The age of secrecy and unfairness in recidivism prediction” In Harvard Data Science Review 2.1, 2020, pp. 1
- “Approximating XGBoost with an interpretable decision tree” In Information Sciences 572 Elsevier, 2021, pp. 522–542
- “Explaining deep neural networks and beyond: A review of methods and applications” In Proceedings of the IEEE 109.3 IEEE, 2021, pp. 247–278
- “Towards explainable artificial intelligence” In Explainable AI: interpreting, explaining and visualizing deep learning Springer, 2019, pp. 5–22
- Nina Schaaf, Marco Huber and Johannes Maucher “Enhancing decision tree based interpretation of deep neural networks through l1-orthogonal regularization” In 2019 18th IEEE International Conference On Machine Learning And Applications (ICMLA), 2019, pp. 42–49 IEEE
- “The relationship between trust in AI and trustworthy machine learning technologies” In Proceedings of the 2020 conference on fairness, accountability, and transparency, 2020, pp. 272–283
- Neil Vigdor “Apple Card Investigated After Gender Discrimination Complaints” In New York Times, 2019 URL: https://www.nytimes.com/2019/11/10/business/Apple-credit-card-investigation.html
- “NBDT: neural-backed decision trees” In arXiv preprint arXiv:2004.00221, 2020
- “In pursuit of interpretable, fair and accurate machine learning for criminal recidivism prediction” In Journal of Quantitative Criminology 39.2 Springer, 2023, pp. 519–581
- Linda Whiteman “The Scale and Effects of Admissions Preferences in Higher Education (SEAPHE)”, 1998 URL: http://www.seaphe.org/databases.php
- W.Bradford Wilcox “Less Poverty, Less Prison, More College: What Two Parents Mean For Black and White Children”, 2021 URL: https://ifstudies.org/blog/less-poverty-less-prison-more-college-what-two-parents-mean-for-black-and-white-children
- Yongxin Yang, Irene Garcia Morillo and Timothy M Hospedales “Deep neural decision trees” In arXiv preprint arXiv:1806.06988, 2018
- “Learning fair representations” In International conference on machine learning, 2013, pp. 325–333 URL: https://proceedings.mlr.press/v28/zemel13.html
- Brian Hu Zhang, Blake Lemoine and Margaret Mitchell “Mitigating Unwanted Biases with Adversarial Learning” In AIES 2018: Proceedings of the 2018 AAAI/ACM Conference on AI, Ethics, and Society, 2018, pp. 335–340 DOI: 10.1145/3278721.3278779
- Indre Zliobaite “On the relation between accuracy and fairness in binary classification” Presented at the 2nd Workshop on Fairness, Accountability, and Transparency in Machine Learning In ArXiv Pre-Print 1505.05723, 2015 DOI: 10.48550/arXiv.1505.05723
- Scott M Lundberg and Su-In Lee “A Unified Approach to Interpreting Model Predictions” In Advances in Neural Information Processing Systems 30 Curran Associates, Inc., 2017 URL: https://proceedings.neurips.cc/paper/2017/file/8a20a8621978632d76c43dfd28b67767-Paper.pdf
- “Layer-wise relevance propagation: an overview” In Explainable AI: interpreting, explaining and visualizing deep learning Springer, 2019, pp. 193–209