Arbitrariness Lies Beyond the Fairness-Accuracy Frontier (2306.09425v1)
Abstract: Machine learning tasks may admit multiple competing models that achieve similar performance yet produce conflicting outputs for individual samples -- a phenomenon known as predictive multiplicity. We demonstrate that fairness interventions in machine learning optimized solely for group fairness and accuracy can exacerbate predictive multiplicity. Consequently, state-of-the-art fairness interventions can mask high predictive multiplicity behind favorable group fairness and accuracy metrics. We argue that a third axis of ``arbitrariness'' should be considered when deploying models to aid decision-making in applications of individual-level impact. To address this challenge, we propose an ensemble algorithm applicable to any fairness intervention that provably ensures more consistent predictions.
- A reductions approach to fair classification. In International Conference on Machine Learning, pages 60–69. PMLR.
- Beyond adult and compas: Fair multi-class prediction via information projection. Advances in Neural Information Processing Systems, 35:38747–38760.
- Ai fairness 360: An extensible toolkit for detecting and mitigating algorithmic bias. IBM Journal of Research and Development, 63(4/5):4–1.
- Selective ensembles for consistent predictions. arXiv preprint arXiv:2111.08230.
- Model multiplicity: Opportunities, concerns, and solutions. In 2022 ACM Conference on Fairness, Accountability, and Transparency, pages 850–863.
- Breiman, L. (1996). Bagging predictors. Machine learning, 24:123–140.
- Breiman, L. (2001). Statistical modeling: The two cultures (with comments and a rejoinder by the author). Statistical science, 16(3):199–231.
- Fairness in machine learning: A survey. arXiv preprint arXiv:2010.04053.
- Chouldechova, A. (2017). Fair prediction with disparate impact: A study of bias in recidivism prediction instruments. Big data, 5(2):153–163.
- Leveraging labeled and unlabeled data for consistent fair binary classification. Advances in Neural Information Processing Systems, 32.
- Variance, self-consistency, and arbitrariness in fair classification. arXiv preprint arXiv:2301.11562.
- Characterizing fairness over the set of good models under selective labels. In International Conference on Machine Learning, pages 2144–2155. PMLR.
- The algorithmic leviathan: arbitrariness, fairness, and opportunity in algorithmic decision-making systems. Canadian Journal of Philosophy, 52(1):26–43.
- Cury, C. R. J. (2022). Instituto nacional de estudos e pesquisas educacionais anísio teixeira: uma trajetória em busca de uma educação de qualidade.
- Underspecification presents challenges for credibility in modern machine learning. The Journal of Machine Learning Research, 23(1):10237–10297.
- Preserving statistical validity in adaptive data analysis. In Proceedings of the forty-seventh annual ACM symposium on Theory of computing, pages 117–126.
- Fairness through awareness. In Proceedings of the 3rd Innovations in Theoretical Computer Science Conference, ITCS ’12, page 214–226, New York, NY, USA. Association for Computing Machinery.
- All models are wrong, but many are useful: Learning a variable’s importance by studying an entire class of prediction models simultaneously. J. Mach. Learn. Res., 20(177):1–81.
- On the impact of machine learning randomness on group fairness.
- Statistical learning theory. Lecture Notes, 387.
- Equality of opportunity in supervised learning. Advances in neural information processing systems, 29.
- Multicalibration: Calibration for the (computationally-identifiable) masses. In International Conference on Machine Learning, pages 1939–1948. PMLR.
- Bia mitigation for machine learning classifiers: A comprehensive survey. arXiv preprint arXiv:2207.07068.
- Rashomon capacity: A metric for predictive multiplicity in probabilistic classification. arXiv preprint arXiv:2206.01295.
- High school longitudinal study of 2009 (hsls: 09): Base-year data file documentation. nces 2011-328. National Center for Education Statistics.
- Fairness without imputation: A decision tree approach for fair prediction with missing values. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 36, pages 9558–9566.
- Decision theory for discrimination-aware classification. In 2012 IEEE 12th international conference on data mining, pages 924–929. IEEE.
- Multiaccuracy: Black-box post-processing for fairness in classification. In Proceedings of the 2019 AAAI/ACM Conference on AI, Ethics, and Society, pages 247–254.
- When mitigating bias is unfair: A comprehensive study on the impact of bias mitigation algorithms. arXiv preprint arXiv:2302.07185.
- Arbitrary decisions are a hidden cost of differentially-private training. In Conference on Fairness, Accountability, and Transparency (FAccT). ACM.
- Lichman, M. et al. (2013). Uci machine learning repository, 2013.
- Predictive multiplicity in classification. In International Conference on Machine Learning, pages 6765–6774. PMLR.
- Scikit-learn: Machine learning in python. the Journal of machine Learning research, 12:2825–2830.
- On fairness and calibration. Advances in neural information processing systems, 30.
- Reconciling individual probability forecasts. arXiv preprint arXiv:2209.01687.
- A study in rashomon curves and volumes: A new perspective on generalization and model simplicity in machine learning. arXiv preprint arXiv:1908.01755.
- Predictive multiplicity in probabilistic classification. arXiv preprint arXiv:2206.01131.
- Data mining: practical machine learning tools and techniques with java implementations. Acm Sigmod Record, 31(1):76–77.
- Carol Xuan Long (4 papers)
- Hsiang Hsu (24 papers)
- Wael Alghamdi (8 papers)
- Flavio P. Calmon (56 papers)