Mind the Gap: A Causal Perspective on Bias Amplification in Prediction & Decision-Making (2405.15446v1)
Abstract: Investigating fairness and equity of automated systems has become a critical field of inquiry. Most of the literature in fair machine learning focuses on defining and achieving fairness criteria in the context of prediction, while not explicitly focusing on how these predictions may be used later on in the pipeline. For instance, if commonly used criteria, such as independence or sufficiency, are satisfied for a prediction score $S$ used for binary classification, they need not be satisfied after an application of a simple thresholding operation on $S$ (as commonly used in practice). In this paper, we take an important step to address this issue in numerous statistical and causal notions of fairness. We introduce the notion of a margin complement, which measures how much a prediction score $S$ changes due to a thresholding operation. We then demonstrate that the marginal difference in the optimal 0/1 predictor $\widehat Y$ between groups, written $P(\hat y \mid x_1) - P(\hat y \mid x_0)$, can be causally decomposed into the influences of $X$ on the $L_2$-optimal prediction score $S$ and the influences of $X$ on the margin complement $M$, along different causal pathways (direct, indirect, spurious). We then show that under suitable causal assumptions, the influences of $X$ on the prediction score $S$ are equal to the influences of $X$ on the true outcome $Y$. This yields a new decomposition of the disparity in the predictor $\widehat Y$ that allows us to disentangle causal differences inherited from the true outcome $Y$ that exists in the real world vs. those coming from the optimization procedure itself. This observation highlights the need for more regulatory oversight due to the potential for bias amplification, and to address this issue we introduce new notions of weak and strong business necessity, together with an algorithm for assessing whether these notions are satisfied.
- Elston v. Talladega County Bd. of Educ. 997 F.2d 1394 (11th Cir. 1993), 1993. United States Court of Appeals for the Eleventh Circuit.
- C. R. Act. Civil rights act of 1964. Title VII, Equal Employment Opportunities, 1964.
- Causal effect identification in cluster dags. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 37(10), pages 12172–12179, 2023.
- Machine bias: There’s software used across the country to predict future criminals. and it’s biased against blacks. ProPublica, 5 2016. URL https://www.propublica.org/article/machine-bias-risk-assessments-in-criminal-sentencing.
- On pearl’s hierarchy and the foundations of causal inference. In Probabilistic and Causal Inference: The Works of Judea Pearl, page 507–556. Association for Computing Machinery, New York, NY, USA, 1st edition, 2022.
- The gender earnings gap: learning from international comparisons. The American Economic Review, 82(2):533–538, 1992.
- The gender wage gap: Extent, trends, and explanations. Journal of economic literature, 55(3):789–865, 2017.
- Evaluating the predictive validity of the compas risk and needs assessment system. Criminal Justice and Behavior, 36(1):21–40, 2009.
- J. Buolamwini and T. Gebru. Gender shades: Intersectional accuracy disparities in commercial gender classification. In S. A. Friedler and C. Wilson, editors, Proceedings of the 1st Conference on Fairness, Accountability and Transparency, volume 81 of Proceedings of Machine Learning Research, pages 77–91, NY, USA, 2018.
- S. Chiappa. Path-specific counterfactual fairness. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 33, pages 7801–7808, 2019.
- A. Chouldechova. Fair prediction with disparate impact: A study of bias in recidivism prediction instruments. Technical Report arXiv:1703.00056, arXiv.org, 2017.
- R. B. Darlington. Another look at “cultural fairness” 1. Journal of educational measurement, 8(2):71–82, 1971.
- Automated experiments on ad privacy settings: A tale of opacity, choice, and discrimination. Proceedings on Privacy Enhancing Technologies, 2015(1):92–112, Apr. 2015. doi: 10.1515/popets-2015-0007.
- Experimental evaluation of algorithm-assisted human decision-making: Application to pretrial public safety assessment. Journal of the Royal Statistical Society Series A: Statistics in Society, 186(2):167–189, 2023.
- Mimic-iv, a freely accessible electronic health record dataset. Scientific data, 10(1):1, 2023.
- Consumer credit-risk models via machine-learning algorithms. Journal of Banking & Finance, 34(11):2767–2787, 2010.
- Avoiding discrimination through causal reasoning. arXiv preprint arXiv:1706.02744, 2017.
- Human decisions and machine predictions. The quarterly journal of economics, 133(1):237–293, 2018.
- Counterfactual fairness. Advances in neural information processing systems, 30, 2017.
- How we analyzed the compas recidivism algorithm. ProPublica (5 2016), 9, 2016.
- Method and system for loan origination and underwriting, Oct. 23 2007. US Patent 7,287,008.
- R. Nabi and I. Shpitser. Fair inference on outcomes. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 32, 2018.
- Causal conceptions of fairness and their consequences. In International Conference on Machine Learning, pages 16848–16887. PMLR, 2022.
- D. Pager. The mark of a criminal record. American journal of sociology, 108(5):937–975, 2003.
- J. Pearl. Causality: Models, Reasoning, and Inference. Cambridge University Press, New York, 2000. 2nd edition, 2009.
- An algorithmic approach to reducing unexplained pain disparities in underserved populations. Nature Medicine, 27(1):136–140, 2021.
- D. Plečko and E. Bareinboim. Causal fairness analysis: A causal toolkit for fair machine learning. Foundations and Trends® in Machine Learning, 17(3):304–589, 2024.
- D. Plecko and E. Bareinboim. Causal fairness for outcome control. Advances in Neural Information Processing Systems, 36, 2024a.
- D. Plecko and E. Bareinboim. Reconciling predictive and statistical parity: A causal approach. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 38 (13), pages 14625–14632, 2024b.
- D. Plečko and N. Meinshausen. Fair data adaptation with quantile preservation. Journal of Machine Learning Research, 21:242, 2020.
- J. Sanburn. Facebook thinks some native american names are inauthentic. Time, Feb. 14 2015. URL http://time.com/3710203/facebook-native-american-names/.
- I. Shpitser and J. Pearl. What counterfactuals can be tested. In Proceedings of the Twenty-third Conference on Uncertainty in Artificial Intelligence, page 352–359, 2007.
- L. Sweeney. Discrimination in online ad delivery. Technical Report 2208240, SSRN, Jan. 28 2013. URL http://dx.doi.org/10.2139/ssrn.2208240.
- L. T. Sweeney and C. Haney. The influence of race on sentencing: A meta-analytic review of experimental studies. Behavioral Sciences & the Law, 10(2):179–195, 1992.
- Pc-fairness: A unified framework for measuring causality-based fairness. Advances in neural information processing systems, 32, 2019.
- J. Zhang and E. Bareinboim. Equality of opportunity in classification: A causal approach. In S. Bengio, H. Wallach, H. Larochelle, K. Grauman, N. Cesa-Bianchi, and R. Garnett, editors, Advances in Neural Information Processing Systems 31, pages 3671–3681, Montreal, Canada, 2018a. Curran Associates, Inc.
- J. Zhang and E. Bareinboim. Fairness in decision-making—the causal explanation formula. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 32, 2018b.
- Partial counterfactual identification from observational and experimental data. In Proceedings of the 39th International Conference on Machine Learning, 2022.
- Drago Plecko (12 papers)
- Elias Bareinboim (34 papers)