Auditing Predictive Models for Intersectional Biases (2306.13064v1)
Abstract: Predictive models that satisfy group fairness criteria in aggregate for members of a protected class, but do not guarantee subgroup fairness, could produce biased predictions for individuals at the intersection of two or more protected classes. To address this risk, we propose Conditional Bias Scan (CBS), a flexible auditing framework for detecting intersectional biases in classification models. CBS identifies the subgroup for which there is the most significant bias against the protected class, as compared to the equivalent subgroup in the non-protected class, and can incorporate multiple commonly used fairness definitions for both probabilistic and binarized predictions. We show that this methodology can detect previously unidentified intersectional and contextual biases in the COMPAS pre-trial risk assessment tool and has higher bias detection power compared to similar methods that audit for subgroup fairness.
- Machine bias. https://www.propublica.org/article/machine-bias-risk-assessments-in-criminal-sentencing
- Machine bias. In Ethics of Data and Analytics. Auerbach Publications, 254–264.
- Vance W Berger. 2000. Pros and cons of permutation tests in clinical trials. Statistics in medicine 19, 10 (2000), 1319–1328.
- Alexandra Chouldechova. 2017. Fair prediction with disparate impact: A study of bias in recidivism prediction instruments. Big data 5, 2 (2017), 153–163.
- Fairness through awareness. In Proceedings of the 3rd innovations in theoretical computer science conference. 214–226.
- Ben Green. 2020. The false promise of risk assessments: epistemic reform and the limits of fairness. In Proceedings of the 2020 conference on fairness, accountability, and transparency. 594–606.
- Guido W Imbens. 2004. Nonparametric estimation of average treatment effects under exogeneity: A review. Review of Economics and statistics 86, 1 (2004), 4–29.
- Preventing fairness gerrymandering: Auditing and learning for subgroup fairness. In International Conference on Machine Learning. PMLR, 2564–2572.
- Multiaccuracy: Black-box post-processing for fairness in classification. In Proceedings of the 2019 AAAI/ACM Conference on AI, Ethics, and Society. ACM, 247–254.
- MultiAccuracyBoost. https://github.com/amiratag/MultiAccuracyBoost.
- Martin Kulldorff. 1997. A spatial scan statistic. Communications in Statistics-Theory and methods 26, 6 (1997), 1481–1496.
- Counterfactual fairness. Advances in neural information processing systems 30 (2017).
- How we analyzed the compas recidivism algorithm. https://www.propublica.org/article/how-we-analyzed-the-compas-recidivism-algorithm
- Jeff Larson and Marjorie Roswell. 2017. Compas-analysis/compas analysis.ipynb at master · PROPUBLICA/Compas-analysis. https://github.com/propublica/compas-analysis/blob/master/Compas%20Analysis.ipynb
- Efficient discovery of heterogeneous treatment effects in randomized experiments via anomalous pattern detection. arXiv preprint arXiv:1803.09159 (2018).
- Flipping the Script on Criminal Justice Risk Assessment: An actuarial model for assessing the risk the federal sentencing system poses to defendants. arXiv preprint arXiv:2205.13505 (2022).
- Algorithmic fairness: Choices, assumptions, and definitions. Annual Review of Statistics and Its Application 8 (2021), 141–163.
- GerryFair: Auditing and Learning for Subgroup Fairness. https://github.com/algowatchpenn/GerryFair.
- Daniel B Neill. 2012. Fast subset scan for spatial pattern detection. Journal of the Royal Statistical Society: Series B (Statistical Methodology) 74, 2 (2012), 337–360.
- Dissecting racial bias in an algorithm used to manage the health of populations. Science 366, 6464 (2019), 447–453.
- Racial differences in marijuana-users’ risk of arrest in the United States. Drug and alcohol dependence 84, 3 (2006), 264–272.
- The age of secrecy and unfairness in recidivism prediction. Harvard Data Science Review 2, 1 (2020), 1.
- Anne Sisson Runyan. 2018. What Is Intersectionality and Why Is It Important? Academe 104, 6 (2018), 10–14.
- Megan S Schuler and Sherri Rose. 2017. Targeted maximum likelihood estimation for causal inference in observational studies. American journal of epidemiology 185, 1 (2017), 65–73.
- Penalized fast subset scanning. Journal of Computational and Graphical Statistics 25, 2 (2016), 382–404.
- Zhe Zhang and Daniel B Neill. 2016. Identifying significant predictive bias in classifiers. arXiv preprint arXiv:1611.08292 (2016).
Sponsor
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.