Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
175 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Fair Feature Importance Scores for Interpreting Tree-Based Methods and Surrogates (2310.04352v1)

Published 6 Oct 2023 in stat.ML and cs.LG

Abstract: Across various sectors such as healthcare, criminal justice, national security, finance, and technology, large-scale ML and AI systems are being deployed to make critical data-driven decisions. Many have asked if we can and should trust these ML systems to be making these decisions. Two critical components are prerequisites for trust in ML systems: interpretability, or the ability to understand why the ML system makes the decisions it does, and fairness, which ensures that ML systems do not exhibit bias against certain individuals or groups. Both interpretability and fairness are important and have separately received abundant attention in the ML literature, but so far, there have been very few methods developed to directly interpret models with regard to their fairness. In this paper, we focus on arguably the most popular type of ML interpretation: feature importance scores. Inspired by the use of decision trees in knowledge distillation, we propose to leverage trees as interpretable surrogates for complex black-box ML models. Specifically, we develop a novel fair feature importance score for trees that can be used to interpret how each feature contributes to fairness or bias in trees, tree-based ensembles, or tree-based surrogates of any complex ML system. Like the popular mean decrease in impurity for trees, our Fair Feature Importance Score is defined based on the mean decrease (or increase) in group bias. Through simulations as well as real examples on benchmark fairness datasets, we demonstrate that our Fair Feature Importance Score offers valid interpretations for both tree-based ensembles and tree-based surrogates of other ML systems.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (53)
  1. “A Reductions Approach to Fair Classification” In ICML 2018: Proceedings of the 35th International Conference on Machine Learning 80, 2018, pp. 60–69 URL: https://proceedings.mlr.press/v80/agarwal18a.html
  2. Sushant Agarwal “Trade-offs between fairness and interpretability in machine learning” In IJCAI 2021 Workshop on AI for Social Good, 2021
  3. Genevera I Allen, Luqin Gan and Lili Zheng “Interpretable Machine Learning for Discovery: Statistical Challenges\\\backslash\& Opportunities” In arXiv preprint arXiv:2308.01475, 2023
  4. “Explainability for fair machine learning” In arXiv preprint arXiv:2010.07389, 2020
  5. “Data Decisions and Theoretical Implications when Adversarially Learning Fair Representations” In ArXiv Pre-Print 1707.00075, 2017 DOI: 10.48550/arXiv.1707.00075
  6. Vaishnavi Bhargava, Miguel Couceiro and Amedeo Napoli “LimeOut: an ensemble approach to improve process fairness” In Joint European conference on machine learning and knowledge discovery in databases, 2020, pp. 475–491 Springer
  7. “Machine learning explainability through comprehensible decision trees” In Machine Learning and Knowledge Extraction: Third IFIP TC 5, TC 12, WG 8.4, WG 8.9, WG 12.9 International Cross-Domain Conference, CD-MAKE 2019, Canterbury, UK, August 26–29, 2019, Proceedings 3, 2019, pp. 15–26 Springer
  8. Leo Breiman “Bagging Predictors” In Machine Learning 24.2 Springer, 1996, pp. 123–140 DOI: 10.1023/A:1018054314350
  9. Leo Breiman “Classification and regression trees” Routledge, 1973
  10. “Toward a taxonomy of trust for probabilistic machine learning” In Science Advances 9.7 American Association for the Advancement of Science, 2023, pp. eabn3999
  11. Cristian Buciluǎ, Rich Caruana and Alexandru Niculescu-Mizil “Model compression” In Proceedings of the 12th ACM SIGKDD international conference on Knowledge discovery and data mining, 2006, pp. 535–541
  12. “Optimized pre-processing for discrimination prevention” In NeurIPS 2017: Advances in Neural Information Processing Systems 30, 2017, pp. 3995–4004 DOI: https://proceedings.neurips.cc/paper/2017/hash/9a49a25d845a483fae4be7e341368e36-Abstract.html
  13. Ian Carlos Campbell “The Apple Card Doesn’t Actually Discriminate Against Women, Investigators Say” In The Verge, 2021 URL: https://www.theverge.com/2021/3/23/22347127/goldman-sachs-apple-card-no-gender-discrimination
  14. “Fairness in machine learning: A survey” In ArXiv Pre-Print 2010.04053, 2020 DOI: 10.48550/arXiv.2010.04053
  15. “L-shapley and c-shapley: Efficient model interpretation for structured data” In arXiv preprint arXiv:1808.02610, 2018
  16. “The Frontiers of Fairness in Machine Learning” In ArXiv Pre-Print 1810.08810, 2018 DOI: 10.48550/arXiv.1810.08810
  17. Mengnan Du, Ninghao Liu and Xia Hu “Techniques for interpretable machine learning” In Communications of the ACM 63.1 ACM New York, NY, USA, 2019, pp. 68–77
  18. “UCI Machine Learning Repository”, 2017 URL: http://archive.ics.uci.edu/ml
  19. “A Comparative Study of Fairness-Enhancing Interventions in Machine Learning” In FAccT 2019: Proceedings of the 2019 Conference on Fairness, Accountability, and Transparency, 2019, pp. 329–338 DOI: 10.1145/3287560.3287589
  20. “Knowledge distillation: A survey” In International Journal of Computer Vision 129 Springer, 2021, pp. 1789–1819
  21. “Fair Adversarial Gradient Tree Boosting” In ICDM 2019: Proceedings of the 2019 IEEE International Conference on Data Mining, 2019, pp. 1060–1065 DOI: 10.1109/ICDM.2019.00124
  22. “A Survey of Methods for Explaining Black Box Models” In ACM Computing Surveys 51.5 New York, NY, USA: Association for Computing Machinery, 2018 DOI: 10.1145/3236009
  23. Moritz Hardt, Eric Price and Nathan Srebron “Equality of Opportunity in Supervised Learning” In NeurIPS 2016: Advances in Neural Information Processing Systems 29 29, 2016 URL: https://papers.nips.cc/paper/2016/hash/9d2682367c3935defcb1f9e247a97c0d-Abstract.html
  24. Geoffrey Hinton, Oriol Vinyals and Jeff Dean “Distilling the knowledge in a neural network” In arXiv preprint arXiv:1503.02531, 2015
  25. Will Knight “The Apple Card Didn’t See Gender—and That’s the Problem” In WIRED, 2019 URL: https://www.wired.com/story/the-apple-card-didnt-see-genderand-thats-the-problem/
  26. Jeff Larson, Marjorie Roswell and Vaggelis Atlidakis “COMPAS Recidivism Risk Score Data and Analysis”, 2022 URL: https://github.com/propublica/compas-analysis/
  27. Zachary C Lipton “The mythos of model interpretability: In machine learning, the concept of interpretability is both important and slippery.” In Queue 16.3, 2018, pp. 31–57
  28. Camille Olivia Little, Michael Weylandt and Genevera I. Allen “To the Fairness Frontier and Beyond: Identifying, Quantifying, and Optimizing the Fairness-Accuracy Pareto Frontier” In arXiv preprint arXiv:2206.00074, 2022
  29. Wei-Yin Loh “Classification and regression trees” In Wiley interdisciplinary reviews: data mining and knowledge discovery Wiley Online Library, 2011, pp. 14–23
  30. “Bias Mitigation Post-processing for Individual and Group Fairness” In ICASSP 2019: Proceedings of the 2019 IEEE International Conference on Acoustics, Speech and Signal Processing, 2019, pp. 2847–2851 DOI: 10.1109/ICASSP.2019.8682620
  31. “Understanding variable importances in forests of randomized trees” In Advances in neural information processing systems 26, 2013
  32. Masayoshi Mase, Art B Owen and Benjamin B Seiler “Cohort Shapley value for algorithmic fairness” In arXiv preprint arXiv:2105.07168, 2021
  33. Christoph Molnar “Interpretable machine learning” 2 edn., 2020
  34. “Definitions, methods, and applications in interpretable machine learning” In Proceedings of the National Academy of Sciences 116.44, 2019, pp. 22071–22080
  35. “Explainable, trustworthy, and ethical machine learning for healthcare: A survey” In Computers in Biology and Medicine Elsevier, 2022, pp. 106043
  36. Marco Tulio Ribeiro, Sameer Singh and Carlos Guestrin “"Why Should I Trust You?": Explaining the Predictions of Any Classifier” In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining Association for Computing Machinery, 2016, pp. 1135–1144 DOI: 10.1145/2939672.2939778
  37. Cynthia Rudin, Caroline Wang and Beau Coker “The age of secrecy and unfairness in recidivism prediction” In Harvard Data Science Review 2.1, 2020, pp. 1
  38. “Approximating XGBoost with an interpretable decision tree” In Information Sciences 572 Elsevier, 2021, pp. 522–542
  39. “Explaining deep neural networks and beyond: A review of methods and applications” In Proceedings of the IEEE 109.3 IEEE, 2021, pp. 247–278
  40. “Towards explainable artificial intelligence” In Explainable AI: interpreting, explaining and visualizing deep learning Springer, 2019, pp. 5–22
  41. Nina Schaaf, Marco Huber and Johannes Maucher “Enhancing decision tree based interpretation of deep neural networks through l1-orthogonal regularization” In 2019 18th IEEE International Conference On Machine Learning And Applications (ICMLA), 2019, pp. 42–49 IEEE
  42. “The relationship between trust in AI and trustworthy machine learning technologies” In Proceedings of the 2020 conference on fairness, accountability, and transparency, 2020, pp. 272–283
  43. Neil Vigdor “Apple Card Investigated After Gender Discrimination Complaints” In New York Times, 2019 URL: https://www.nytimes.com/2019/11/10/business/Apple-credit-card-investigation.html
  44. “NBDT: neural-backed decision trees” In arXiv preprint arXiv:2004.00221, 2020
  45. “In pursuit of interpretable, fair and accurate machine learning for criminal recidivism prediction” In Journal of Quantitative Criminology 39.2 Springer, 2023, pp. 519–581
  46. Linda Whiteman “The Scale and Effects of Admissions Preferences in Higher Education (SEAPHE)”, 1998 URL: http://www.seaphe.org/databases.php
  47. W.Bradford Wilcox “Less Poverty, Less Prison, More College: What Two Parents Mean For Black and White Children”, 2021 URL: https://ifstudies.org/blog/less-poverty-less-prison-more-college-what-two-parents-mean-for-black-and-white-children
  48. Yongxin Yang, Irene Garcia Morillo and Timothy M Hospedales “Deep neural decision trees” In arXiv preprint arXiv:1806.06988, 2018
  49. “Learning fair representations” In International conference on machine learning, 2013, pp. 325–333 URL: https://proceedings.mlr.press/v28/zemel13.html
  50. Brian Hu Zhang, Blake Lemoine and Margaret Mitchell “Mitigating Unwanted Biases with Adversarial Learning” In AIES 2018: Proceedings of the 2018 AAAI/ACM Conference on AI, Ethics, and Society, 2018, pp. 335–340 DOI: 10.1145/3278721.3278779
  51. Indre Zliobaite “On the relation between accuracy and fairness in binary classification” Presented at the 2nd Workshop on Fairness, Accountability, and Transparency in Machine Learning In ArXiv Pre-Print 1505.05723, 2015 DOI: 10.48550/arXiv.1505.05723
  52. Scott M Lundberg and Su-In Lee “A Unified Approach to Interpreting Model Predictions” In Advances in Neural Information Processing Systems 30 Curran Associates, Inc., 2017 URL: https://proceedings.neurips.cc/paper/2017/file/8a20a8621978632d76c43dfd28b67767-Paper.pdf
  53. “Layer-wise relevance propagation: an overview” In Explainable AI: interpreting, explaining and visualizing deep learning Springer, 2019, pp. 193–209
Citations (1)

Summary

We haven't generated a summary for this paper yet.