Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
126 tokens/sec
GPT-4o
47 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Debiasing Machine Learning Models by Using Weakly Supervised Learning (2402.15477v1)

Published 23 Feb 2024 in cs.LG and cs.CY

Abstract: We tackle the problem of bias mitigation of algorithmic decisions in a setting where both the output of the algorithm and the sensitive variable are continuous. Most of prior work deals with discrete sensitive variables, meaning that the biases are measured for subgroups of persons defined by a label, leaving out important algorithmic bias cases, where the sensitive variable is continuous. Typical examples are unfair decisions made with respect to the age or the financial status. In our work, we then propose a bias mitigation strategy for continuous sensitive variables, based on the notion of endogeneity which comes from the field of econometrics. In addition to solve this new problem, our bias mitigation strategy is a weakly supervised learning method which requires that a small portion of the data can be measured in a fair manner. It is model agnostic, in the sense that it does not make any hypothesis on the prediction model. It also makes use of a reasonably large amount of input observations and their corresponding predictions. Only a small fraction of the true output predictions should be known. This therefore limits the need for expert interventions. Results obtained on synthetic data show the effectiveness of our approach for examples as close as possible to real-life applications in econometrics.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (59)
  1. Abien Fred Agarap. Deep learning using rectified linear units (relu). arXiv preprint arXiv:1803.08375, 2018.
  2. Wasserstein generative adversarial networks. In International Conference on Machine Learning, pp.  214–223. PMLR, 2017.
  3. Solving inverse problems using data-driven models. Acta Numerica, 28:1–174, 2019.
  4. Fairness in machine learning. Nips tutorial, 1:2017, 2017.
  5. Fairness and machine learning: Limitations and opportunities. MIT Press, 2023.
  6. Simplicity bias leads to amplified performance disparities. In Proceedings of the 2023 ACM Conference on Fairness, Accountability, and Transparency, pp.  355–369, 2023.
  7. Fairness in criminal justice risk assessments: The state of the art. Sociological Methods & Research, 50(1):3–44, 2021.
  8. A survey of bias in machine learning through the prism of statistical parity. The American Statistician, 76(2):188–198, 2022.
  9. Fairlearn: A toolkit for assessing and improving fairness in AI. Technical report, Microsoft, May 2020. URL https://www.microsoft.com/en-us/research/publication/fairlearn-a-toolkit-for-assessing-and-improving-fairness-in-ai/.
  10. Pattern recognition and machine learning, volume 4. Springer, 2006.
  11. Three naive bayes approaches for discrimination-free classification. Data mining and knowledge discovery, 21(2):277–292, 2010.
  12. Optimized pre-processing for discrimination prevention. Advances in neural information processing systems, 30, 2017.
  13. Linear inverse problems in structural econometrics estimation based on spectral decomposition and regularization. Handbook of econometrics, 6:5633–5751, 2007.
  14. Asymptotic normal inference in linear inverse problems. In J. Racine, L. Su, and A. Ullah (eds.), Handbook of Non Parametric Statistics, pp.  65–96, Oxford, 2014.
  15. Nonparametric estimation of accelerated failure-time models with unobservable confounders and random censoring. Electronic Journal of Statistics, 15(2):5333–5379, 2021.
  16. Fairness in machine learning and econometrics. In Econometrics with Machine Learning, pp.  217–250. Springer, 2022.
  17. A snapshot of the frontiers of fairness in machine learning. Communications of the ACM, 63(5):82–89, 2020.
  18. Marco Cuturi. Sinkhorn distances: Lightspeed computation of optimal transport. In C.J. Burges, L. Bottou, M. Welling, Z. Ghahramani, and K.Q. Weinberger (eds.), Advances in Neural Information Processing Systems, volume 26. Curran Associates, Inc., 2013. URL https://proceedings.neurips.cc/paper_files/paper/2013/file/af21d0c97db2e27e13572cbf59eb343d-Paper.pdf.
  19. A central limit theorem for lp transportation cost on the real line with application to fairness assessment in machine learning. Information and Inference: A Journal of the IMA, 8(4):817–849, 2019.
  20. Review of mathematical frameworks for fairness in machine learning. arXiv preprint arXiv:2005.13755, 2020.
  21. Fairness through awareness. In Proceedings of the 3rd Innovations in Theoretical Computer Science Conference, pp.  214–226, 2012.
  22. Regularization of inverse problems, volume 375. Springer Science & Business Media, 1996.
  23. Certifying and removing disparate impact. In Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp.  259–268, 2015a.
  24. Certifying and removing disparate impact. In proceedings of the 21th ACM SIGKDD international conference on knowledge discovery and data mining, pp.  259–268, 2015b.
  25. Interpolating between optimal transport and mmd using sinkhorn divergences. In The 22nd International Conference on Artificial Intelligence and Statistics, pp.  2681–2690, 2019.
  26. Jean-Pierre Florens. Inverse problems and structural econometrics: The example of instrumental variables. In M. Dewatripont, L.P. Hansen, and S. Turnosky (eds.), Advances Economics and Econometrics, Theory and Applications, pp.  284–311, Cambridge, UK, 2003. Cambridge University Press.
  27. The (Im)possibility of fairness: different value systems require different mechanisms for fair decision making. Communications of the ACM, 64(4):136–143, March 2021. ISSN 0001-0782. doi: 10.1145/3433949. URL https://dl.acm.org/doi/10.1145/3433949.
  28. Learning with a wasserstein loss. Advances in Neural Information Processing Systems, 28, 2015.
  29. Solving inverse problems with deep neural networks–robustness included? IEEE Transactions on Pattern Analysis and Machine Intelligence, 45(1):1119–1134, 2022.
  30. Deep Learning. MIT Press, Cambridge, MA, USA, 2016.
  31. Obtaining fairness using optimal transport theory. In International Conference on Machine Learning, pp.  2357–2365, 2019.
  32. Equality of opportunity in supervised learning. In D. Lee, M. Sugiyama, U. Luxburg, I. Guyon, and R. Garnett (eds.), Advances in Neural Information Processing Systems, volume 29. Curran Associates, Inc., 2016. URL https://proceedings.neurips.cc/paper/2016/file/9d2682367c3935defcb1f9e247a97c0d-Paper.pdf.
  33. Wasserstein-based projections with applications to inverse problems. SIAM Journal on Mathematics of Data Science, 4(2):581–603, 2022.
  34. Wasserstein fair classification. In Uncertainty in artificial intelligence, pp.  862–872. PMLR, 2020.
  35. Data preprocessing techniques for classification without discrimination. Knowledge and information systems, 33(1):1–33, 2012.
  36. Discrimination aware decision tree learning. In 2010 IEEE International Conference on Data Mining, pp.  869–874. IEEE, 2010.
  37. Fairness-aware classifier with prejudice remover regularizer. In Joint European conference on machine learning and knowledge discovery in databases, pp.  35–50. Springer, 2012.
  38. Adam: A method for stochastic optimization, 2017.
  39. Introducing the expohedron for efficient pareto-optimal fairness-utility amortizations in repeated rankings. In Proceedings of the Fifteenth ACM International Conference on Web Search and Data Mining, pp.  498–507, 2022.
  40. Counterfactual fairness. Advances in Neural Information Processing Systems, 30, 2017.
  41. Adaptive complexity regularization for linear inverse problems. Electronic Journal of Statistics, 2:661–677, 2008.
  42. Adaptive estimation for an inverse regression model with unknown operator. Statistics & Risk Modeling, 29(3):215–242, 2012.
  43. Fairness-aware learning for continuous attributes and treatments. In International Conference on Machine Learning, pp.  4382–4391. PMLR, 2019.
  44. Microeconomic theory, volume 1. Oxford university press New York, 1995.
  45. A survey on bias and fairness in machine learning. ACM Computing Surveys (CSUR), 54(6):1–35, 2021.
  46. Algorithm unrolling: Interpretable, efficient deep learning for signal and image processing. IEEE Signal Processing Magazine, 38(2):18–44, 2021.
  47. End-to-end reconstruction meets data-driven regularization for inverse problems. Advances in Neural Information Processing Systems, 34:21413–21425, 2021.
  48. Model specification and endogeneity. Journal of Econometrics, 83(1-2):213–237, 1998.
  49. Fairness in machine learning. In Recent trends in learning from data: Tutorials from the inns big data and deep learning conference (innsbddl2019), pp.  155–196. Springer, 2020.
  50. General fair empirical risk minimization. In 2020 International Joint Conference on Neural Networks (IJCNN), pp.  1–8. IEEE, 2020.
  51. Automatic differentiation in pytorch. 2017.
  52. A survey of identification and mitigation of machine learning algorithmic biases in image analysis. arXiv preprint arXiv:2210.04491, 2022a.
  53. Tackling algorithmic bias in neural-network classifiers using wasserstein-2 regularization. Journal of Mathematical Imaging and Vision, 64(6):672–689, 2022b.
  54. The price of fair pca: One extra dimension. In S. Bengio, H. Wallach, H. Larochelle, K. Grauman, N. Cesa-Bianchi, and R. Garnett (eds.), Advances in Neural Information Processing Systems, volume 31. Curran Associates, Inc., 2018. URL https://proceedings.neurips.cc/paper/2018/file/cc4af25fa9d2d5c953496579b75f6f6c-Paper.pdf.
  55. Cédric Villani et al. Optimal transport: old and new, volume 338. Springer, 2009.
  56. A survey on the fairness of recommender systems. ACM Transactions on Information Systems, 41(3):1–43, 2023.
  57. Learning non-discriminatory predictors. In Conference on Learning Theory, pp.  1920–1953. PMLR, 2017.
  58. Fairness constraints: Mechanisms for fair classification. Proceedings of the 20th Artificial Intelligence and Statistics (20-22 April 2017, Fort Lauderdale, FL, USA), 2017.
  59. Zhi-Hua Zhou. A brief introduction to weakly supervised learning. National Science Review, 5(1):44–53, 2018.

Summary

We haven't generated a summary for this paper yet.