Papers
Topics
Authors
Recent
2000 character limit reached

Understanding Fairness Surrogate Functions in Algorithmic Fairness (2310.11211v4)

Published 17 Oct 2023 in cs.LG and cs.AI

Abstract: It has been observed that machine learning algorithms exhibit biased predictions against certain population groups. To mitigate such bias while achieving comparable accuracy, a promising approach is to introduce surrogate functions of the concerned fairness definition and solve a constrained optimization problem. However, it is intriguing in previous work that such fairness surrogate functions may yield unfair results and high instability. In this work, in order to deeply understand them, taking a widely used fairness definition--demographic parity as an example, we show that there is a surrogate-fairness gap between the fairness definition and the fairness surrogate function. Also, the theoretical analysis and experimental results about the gap motivate us that the fairness and stability will be affected by the points far from the decision boundary, which is the large margin points issue investigated in this paper. To address it, we propose the general sigmoid surrogate to simultaneously reduce both the surrogate-fairness gap and the variance, and offer a rigorous fairness and stability upper bound. Interestingly, the theory also provides insights into two important issues that deal with the large margin points as well as obtaining a more balanced dataset are beneficial to fairness and stability. Furthermore, we elaborate a novel and general algorithm called Balanced Surrogate, which iteratively reduces the gap to mitigate unfairness. Finally, we provide empirical evidence showing that our methods consistently improve fairness and stability while maintaining accuracy comparable to the baselines in three real-world datasets.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (78)
  1. Gpt-4 technical report. arXiv preprint arXiv:2303.08774, 2023.
  2. A reductions approach to fair classification. In Proceedings of the 35th International Conference on Machine Learning, 2018.
  3. Big data’s disparate impact. California law review, pp.  671–732, 2016.
  4. Fairness and Machine Learning. fairmlbook.org, 2019. http://www.fairmlbook.org.
  5. Penalizing unfairness in binary classification. arXiv preprint arXiv:1707.00044, 2017.
  6. The possibility of fairness: Revisiting the impossibility theorem in practice. In Proceedings of the 2023 ACM Conference on Fairness, Accountability, and Transparency, pp.  400–422, 2023.
  7. Scalable and stable surrogates for flexible classifiers with fairness constraints. In Advances in Neural Information Processing Systems, 2021.
  8. A convex framework for fair regression. In Fairness, Accountability, and Transparency in Machine Learning (FATML), 2017.
  9. Understanding the origins of bias in word embeddings. In International conference on machine learning, pp.  803–811, 2019.
  10. Building classifiers with independency constraints. In 2009 IEEE International Conference on Data Mining Workshops, 2009.
  11. Optimized pre-processing for discrimination prevention. In Advances in Neural Information Processing Systems, 2017.
  12. Fairness in machine learning: A survey. arXiv preprint arXiv:2010.04053, 2020.
  13. Fairness with adaptive weights. In International Conference on Machine Learning, pp.  2853–2866, 2022.
  14. Smote: synthetic minority over-sampling technique. Journal of artificial intelligence research, 16:321–357, 2002.
  15. Why is my classifier discriminatory? Advances in neural information processing systems, 31, 2018.
  16. Alexandra Chouldechova. Fair prediction with disparate impact: A study of bias in recidivism prediction instruments. Big data, 5(2):153–163, 2017.
  17. Fair mixup: Fairness via interpolation. In Proceedings of the International Conference on Machine Learning, 2021.
  18. Promoting fairness through hyperparameter optimization. In 2021 IEEE International Conference on Data Mining (ICDM), pp.  1036–1041, 2021.
  19. Differentially private and fair classification via calibrated functional mechanism. Proceedings of the AAAI Conference on Artificial Intelligence, 2020.
  20. Empirical risk minimization under fairness constraints. In Advances in Neural Information Processing Systems, 2018.
  21. Is there a trade-off between fairness and accuracy? a perspective using mismatched hypothesis testing. In International conference on machine learning, pp.  2803–2813. PMLR, 2020.
  22. Certifying and removing disparate impact. In Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 2015.
  23. A confidence-based approach for balancing fairness and accuracy. In Proceedings of the 2016 SIAM International Conference on Data Mining, 2016.
  24. A comparative study of fairness-enhancing interventions in machine learning. In Proceedings of the conference on fairness, accountability, and transparency, pp.  329–338, 2019.
  25. Bias and fairness in large language models: A survey. arXiv:2309.00770, 2023.
  26. On the impact of machine learning randomness on group fairness. In Proceedings of the 2023 ACM Conference on Fairness, Accountability, and Transparency, pp.  1789–1800, 2023.
  27. Satisfying real-world goals with dataset constraints. In Advances in Neural Information Processing Systems, 2016.
  28. Equality of opportunity in supervised learning. In Proceedings of the 30th International Conference on Neural Information Processing Systems, 2016.
  29. Bia mitigation for machine learning classifiers: A comprehensive survey. arXiv preprint arXiv:2207.07068, 2022.
  30. Surya Mattu Julia Angwin, Jeff Larson and Lauren Kirchner. Machine bias: There’s software used across the country to predict future criminals. and it’s biased against blacks. propublica, 2016. URL https://www.propublica.org/article/machine-bias-risk-assessments-in-criminal-sentencing.
  31. Data pre-processing techniques for classification without discrimination. Knowledge and Information Systems, 2011.
  32. Decision theory for discrimination-aware classification. In 2012 IEEE 12th International Conference on Data Mining, 2012.
  33. Fairness-aware classifier with prejudice remover regularizer. In Proceedings of the 2012th European Conference on Machine Learning and Knowledge Discovery in Databases - Volume Part II, 2012.
  34. Chatgpt for good? on opportunities and challenges of large language models for education. Learning and individual differences, 103:102274, 2023.
  35. Inherent trade-offs in the fair determination of risk scores. In Innovations in Theoretical Computer Science (ITCS), 2017.
  36. Ron Kohavi. Scaling up the accuracy of naive-bayes classifiers: A decision-tree hybrid. In Proceedings of the Second International Conference on Knowledge Discovery and Data Mining, 1996.
  37. Adaptive sensitive reweighting to mitigate bias in fairness-aware classification. In Proceedings of the 2018 world wide web conference, pp.  853–862, 2018.
  38. Fairness without demographics through adversarially reweighted learning. Advances in neural information processing systems, 33:728–740, 2020.
  39. A survey on fairness in large language models. arXiv preprint arXiv:2308.10149, 2023.
  40. Does mitigating ml’s impact disparity require treatment disparity? Advances in neural information processing systems, 31, 2018.
  41. The implicit fairness criterion of unconstrained learning. In International Conference on Machine Learning, pp.  4051–4060, 2019.
  42. Trustworthy llms: a survey and guideline for evaluating large language models’ alignment. arXiv preprint arXiv:2308.05374, 2023a.
  43. Summary of chatgpt-related research and perspective towards the future of large language models. Meta-Radiology, 2023b.
  44. Too relaxed to be fair. In Proceedings of the 37th International Conference on Machine Learning, 2020.
  45. Learning adversarially fair and transferable representations. In International Conference on Machine Learning, pp.  3384–3393, 2018.
  46. Fairgrad: Fairness aware gradient descent. Transactions on Machine Learning Research, 2023.
  47. Stability and convergence of stochastic gradient clipping: Beyond lipschitz continuity and smoothness. In International Conference on Machine Learning, pp.  7325–7335, 2021.
  48. A survey on bias and fairness in machine learning. ACM Computing Surveys, 54:1–35, 2021.
  49. Evaluating the fairness of deep learning uncertainty estimates in medical image analysis. In Medical Imaging with Deep Learning, pp.  1453–1492. PMLR, 2024.
  50. Foundations of machine learning. MIT press, 2018.
  51. Addressing fairness in classification with a model-agnostic multi-objective algorithm. In Proceedings of the Thirty-Seventh Conference on Uncertainty in Artificial Intelligence, 2021.
  52. Algorithmic fairness. Machine Learning for Data Science Handbook: Data Mining and Knowledge Discovery Handbook, pp.  867–886, 2023.
  53. On fairness and calibration. In Advances in Neural Information Processing Systems, pp.  5684–5693, 2017.
  54. Towards tracing trustworthiness dynamics: Revisiting pre-training period of large language models. arXiv preprint arXiv:2402.19465, 2024.
  55. Do we reach desired disparate impact with in-processing fairness techniques? Procedia Computer Science, 214:257–264, 2022.
  56. Sample selection for fair and robust training. Advances in Neural Information Processing Systems, 34:815–827, 2021a.
  57. Fairbatch: Batch selection for model fairness. In International Conference on Learning Representations, 2021b.
  58. P. Cortez S. Moro and P. Rita. A data-driven approach to predict the success of bank telemarketing, 2014. URL https://archive.ics.uci.edu/ml/datasets/bank+marketing.
  59. Fair representation learning through implicit path alignment. In Proceedings of the 39th International Conference on Machine Learning, volume 162, pp.  20156–20175, 2022a.
  60. On learning fairness and accuracy on multiple subgroups. Advances in Neural Information Processing Systems, 35:34121–34135, 2022b.
  61. Mitigating calibration bias without fixed attribute grouping for improved fairness in medical imaging analysis. In International Conference on Medical Image Computing and Computer-Assisted Intervention, pp.  189–198. Springer, 2023.
  62. Fair scratch tickets: Finding fair sparse networks without weight training. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp.  24406–24416, 2023.
  63. Large language models in medicine. Nature medicine, 29(8):1930–1940, 2023.
  64. Fairness-aware configuration of machine learning libraries. In Proceedings of the 44th International Conference on Software Engineering, pp.  909–920, 2022.
  65. Llama 2: Open foundation and fine-tuned chat models. arXiv preprint arXiv:2307.09288, 2023.
  66. In-processing modeling techniques for machine learning fairness: A survey. ACM Transactions on Knowledge Discovery from Data, 17(3):1–27, 2023.
  67. Decodingtrust: A comprehensive assessment of trustworthiness in gpt models. In Thirty-seventh Conference on Neural Information Processing Systems Datasets and Benchmarks Track, 2023.
  68. On convexity and bounds of fairness-aware classification. In The World Wide Web Conference, 2019.
  69. Fairness beyond disparate treatment &\&& disparate impact. Proceedings of the 26th International Conference on World Wide Web, 2017a.
  70. From parity to preference-based notions of fairness in classification. In Proceedings of the 31st International Conference on Neural Information Processing Systems, 2017b.
  71. Fairness constraints: Mechanisms for fair classification. In Proceedings of the 20th International Conference on Artificial Intelligence and Statistics, 2017c.
  72. Fairness constraints: A flexible approach for fair classification. Journal of Machine Learning Research, 2019.
  73. Learning fair representations. In Proceedings of the 30th International Conference on Machine Learning, 2013.
  74. Data-centric artificial intelligence: A survey. arXiv preprint arXiv:2303.10158, 2023.
  75. Mitigating unwanted biases with adversarial learning. In Proceedings of the 2018 AAAI/ACM Conference on AI, Ethics, and Society, pp.  335–340, 2018.
  76. Fairness constraints in semi-supervised learning. arXiv preprint arXiv:2009.06190, 2020.
  77. Fair meta-learning for few-shot classification. In 2020 IEEE International Conference on Knowledge Graph (ICKG), 2020.
  78. Inherent tradeoffs in learning fair representations. The Journal of Machine Learning Research, 23(1):2527–2552, 2022.
Citations (1)

Summary

We haven't generated a summary for this paper yet.

Slide Deck Streamline Icon: https://streamlinehq.com

Whiteboard

Dice Question Streamline Icon: https://streamlinehq.com

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Lightbulb Streamline Icon: https://streamlinehq.com

Continue Learning

We haven't generated follow-up questions for this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.