Papers
Topics
Authors
Recent
Detailed Answer
Quick Answer
Concise responses based on abstracts only
Detailed Answer
Well-researched responses based on abstracts and relevant paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses
Gemini 2.5 Flash
Gemini 2.5 Flash 60 tok/s
Gemini 2.5 Pro 51 tok/s Pro
GPT-5 Medium 18 tok/s Pro
GPT-5 High 14 tok/s Pro
GPT-4o 77 tok/s Pro
Kimi K2 159 tok/s Pro
GPT OSS 120B 456 tok/s Pro
Claude Sonnet 4 37 tok/s Pro
2000 character limit reached

FairIF: Boosting Fairness in Deep Learning via Influence Functions with Validation Set Sensitive Attributes (2201.05759v2)

Published 15 Jan 2022 in cs.LG, cs.CY, and stat.ML

Abstract: Most fair machine learning methods either highly rely on the sensitive information of the training samples or require a large modification on the target models, which hinders their practical application. To address this issue, we propose a two-stage training algorithm named FAIRIF. It minimizes the loss over the reweighted data set (second stage) where the sample weights are computed to balance the model performance across different demographic groups (first stage). FAIRIF can be applied on a wide range of models trained by stochastic gradient descent without changing the model, while only requiring group annotations on a small validation set to compute sample weights. Theoretically, we show that, in the classification setting, three notions of disparity among different groups can be mitigated by training with the weights. Experiments on synthetic data sets demonstrate that FAIRIF yields models with better fairness-utility trade-offs against various types of bias; and on real-world data sets, we show the effectiveness and scalability of FAIRIF. Moreover, as evidenced by the experiments with pretrained models, FAIRIF is able to alleviate the unfairness issue of pretrained models without hurting their performance.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (61)
  1. One-Network Adversarial Fairness. In The Thirty-Third AAAI Conference on Artificial Intelligence, AAAI 2019, The Thirty-First Innovative Applications of Artificial Intelligence Conference, IAAI 2019, The Ninth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2019, Honolulu, Hawaii, USA, January 27 - February 1, 2019. AAAI Press, 2412–2420. https://doi.org/10.1609/aaai.v33i01.33012412
  2. A Reductions Approach to Fair Classification. In Proc. of ICML (Proceedings of Machine Learning Research, Vol. 80), Jennifer G. Dy and Andreas Krause (Eds.). PMLR, 60–69. http://proceedings.mlr.press/v80/agarwal18a.html
  3. Second-order stochastic optimization in linear time. stat 1050 (2016), 15.
  4. Arthur Asuncion and David Newman. 2007. UCI machine learning repository.
  5. RelatIF: Identifying Explanatory Training Samples via Relative Influence. In The 23rd International Conference on Artificial Intelligence and Statistics, AISTATS 2020, 26-28 August 2020, Online [Palermo, Sicily, Italy] (Proceedings of Machine Learning Research, Vol. 108), Silvia Chiappa and Roberto Calandra (Eds.). PMLR, 1899–1909. http://proceedings.mlr.press/v108/barshan20a.html
  6. Localized Rademacher Complexities. In Computational Learning Theory, 15th Annual Conference on Computational Learning Theory, COLT 2002, Sydney, Australia, July 8-10, 2002, Proceedings (Lecture Notes in Computer Science, Vol. 2375), Jyrki Kivinen and Robert H. Sloan (Eds.). Springer, 44–58. https://doi.org/10.1007/3-540-45435-7_4
  7. Influence Functions in Deep Learning Are Fragile. In Proc. of ICLR. OpenReview.net. https://openreview.net/forum?id=xHKVVHGDOEk
  8. On Second-Order Group Influence Functions for Black-Box Predictions. In Proc. of ICML (Proceedings of Machine Learning Research, Vol. 119). PMLR, 715–724. http://proceedings.mlr.press/v119/basu20b.html
  9. AI Fairness 360: An extensible toolkit for detecting, understanding, and mitigating unwanted algorithmic bias. ArXiv preprint abs/1810.01943 (2018). https://arxiv.org/abs/1810.01943
  10. Coresets via Bilevel Optimization for Continual Learning and Streaming. In Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, NeurIPS 2020, December 6-12, 2020, virtual, Hugo Larochelle, Marc’Aurelio Ranzato, Raia Hadsell, Maria-Florina Balcan, and Hsuan-Tien Lin (Eds.). https://proceedings.neurips.cc/paper/2020/hash/aa2a77371374094fe9e0bc1de3f94ed9-Abstract.html
  11. A clarification of the nuances in the fairness metrics landscape. Scientific Reports 12, 1 (2022), 4209.
  12. Simon Caton and Christian Haas. 2020. Fairness in machine learning: A survey. ArXiv preprint abs/2010.04053 (2020). https://arxiv.org/abs/2010.04053
  13. Bias in Machine Learning Software: Why? How? What to do?. In Proceedings of the 29th ACM Joint Meeting on European Software Engineering Conference and Symposium on the Foundations of Software Engineering.
  14. COMPAS risk scales: Demonstrating accuracy equity and predictive parity. Northpointe Inc 7, 7.4 (2016), 1.
  15. Fairness through awareness. In Proceedings of the 3rd innovations in theoretical computer science conference. 214–226.
  16. All the cool kids, how do they fit in?: Popularity and demographic biases in recommender evaluation and effectiveness. In FAccT. PMLR, 172–186.
  17. Certifying and Removing Disparate Impact. In Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Sydney, NSW, Australia, August 10-13, 2015, Longbing Cao, Chengqi Zhang, Thorsten Joachims, Geoffrey I. Webb, Dragos D. Margineantu, and Graham Williams (Eds.). ACM, 259–268. https://doi.org/10.1145/2783258.2783311
  18. Beyond Distributive Fairness in Algorithmic Decision Making: Feature Selection for Procedurally Fair Learning. In Proc. of AAAI, Sheila A. McIlraith and Kilian Q. Weinberger (Eds.). AAAI Press, 51–60. https://www.aaai.org/ocs/index.php/AAAI/AAAI18/paper/view/16523
  19. Studying Large Language Model Generalization with Influence Functions. arXiv:2308.03296 [cs.LG]
  20. FastIF: Scalable Influence Functions for Efficient Model Interpretation and Debugging. In Proc. of EMNLP. Association for Computational Linguistics, Online and Punta Cana, Dominican Republic, 10333–10350. https://doi.org/10.18653/v1/2021.emnlp-main.808
  21. Algorithmic Bias: From Discrimination Discovery to Fairness-aware Data Mining. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, August 13-17, 2016, Balaji Krishnapuram, Mohak Shah, Alexander J. Smola, Charu C. Aggarwal, Dou Shen, and Rajeev Rastogi (Eds.). ACM, 2125–2126. https://doi.org/10.1145/2939672.2945386
  22. Equality of Opportunity in Supervised Learning. In Advances in Neural Information Processing Systems 29: Annual Conference on Neural Information Processing Systems 2016, December 5-10, 2016, Barcelona, Spain, Daniel D. Lee, Masashi Sugiyama, Ulrike von Luxburg, Isabelle Guyon, and Roman Garnett (Eds.). 3315–3323. https://proceedings.neurips.cc/paper/2016/hash/9d2682367c3935defcb1f9e247a97c0d-Abstract.html
  23. Deep Residual Learning for Image Recognition. In 2016 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2016, Las Vegas, NV, USA, June 27-30, 2016. IEEE Computer Society, 770–778. https://doi.org/10.1109/CVPR.2016.90
  24. Categorical Reparameterization with Gumbel-Softmax. In Proc. of ICLR. OpenReview.net. https://openreview.net/forum?id=rkE3y85ee
  25. Heinrich Jiang and Ofir Nachum. 2020. Identifying and Correcting Label Bias in Machine Learning. In The 23rd International Conference on Artificial Intelligence and Statistics, AISTATS 2020, 26-28 August 2020, Online [Palermo, Sicily, Italy] (Proceedings of Machine Learning Research, Vol. 108), Silvia Chiappa and Roberto Calandra (Eds.). PMLR, 702–712. http://proceedings.mlr.press/v108/jiang20a.html
  26. Kimmo Karkkainen and Jungseock Joo. 2021. FairFace: Face Attribute Dataset for Balanced Race, Gender, and Age for Bias Measurement and Mitigation. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision. 1548–1558.
  27. On the Accuracy of Influence Functions for Measuring Group Effects. In Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, NeurIPS 2019, December 8-14, 2019, Vancouver, BC, Canada, Hanna M. Wallach, Hugo Larochelle, Alina Beygelzimer, Florence d’Alché-Buc, Emily B. Fox, and Roman Garnett (Eds.). 5255–5265. https://proceedings.neurips.cc/paper/2019/hash/a78482ce76496fcf49085f2190e675b4-Abstract.html
  28. Pang Wei Koh and Percy Liang. 2017. Understanding Black-box Predictions via Influence Functions. In Proc. of ICML (Proceedings of Machine Learning Research, Vol. 70), Doina Precup and Yee Whye Teh (Eds.). PMLR, 1885–1894. http://proceedings.mlr.press/v70/koh17a.html
  29. Adaptive Sensitive Reweighting to Mitigate Bias in Fairness-aware Classification. In Proceedings of the 2018 World Wide Web Conference on World Wide Web, WWW 2018, Lyon, France, April 23-27, 2018, Pierre-Antoine Champin, Fabien L. Gandon, Mounia Lalmas, and Panagiotis G. Ipeirotis (Eds.). ACM, 853–862. https://doi.org/10.1145/3178876.3186133
  30. Fairness without Demographics through Adversarially Reweighted Learning. In Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, NeurIPS 2020, December 6-12, 2020, virtual, Hugo Larochelle, Marc’Aurelio Ranzato, Raia Hadsell, Maria-Florina Balcan, and Hsuan-Tien Lin (Eds.). https://proceedings.neurips.cc/paper/2020/hash/07fc15c9d169ee48573edd749d25945d-Abstract.html
  31. Gradient-based learning applied to document recognition. Proc. IEEE 86, 11 (1998), 2278–2324.
  32. Peizhao Li and Hongfu Liu. 2022. Achieving Fairness at No Utility Cost via Data Reweighing with Influence. In International Conference on Machine Learning, ICML 2022, 17-23 July 2022, Baltimore, Maryland, USA (Proceedings of Machine Learning Research, Vol. 162), Kamalika Chaudhuri, Stefanie Jegelka, Le Song, Csaba Szepesvári, Gang Niu, and Sivan Sabato (Eds.). PMLR, 12917–12930. https://proceedings.mlr.press/v162/li22p.html
  33. Just Train Twice: Improving Group Robustness without Training Group Information. In Proc. of ICML (Proceedings of Machine Learning Research, Vol. 139), Marina Meila and Tong Zhang (Eds.). PMLR, 6781–6792. http://proceedings.mlr.press/v139/liu21f.html
  34. Deep Learning Face Attributes in the Wild. In 2015 IEEE International Conference on Computer Vision, ICCV 2015, Santiago, Chile, December 7-13, 2015. IEEE Computer Society, 3730–3738. https://doi.org/10.1109/ICCV.2015.425
  35. Learning Adversarially Fair and Transferable Representations. In Proc. of ICML (Proceedings of Machine Learning Research, Vol. 80), Jennifer G. Dy and Andreas Krause (Eds.). PMLR, 3381–3390. http://proceedings.mlr.press/v80/madras18a.html
  36. A survey on bias and fairness in machine learning. ACM Computing Surveys (CSUR) 54, 6 (2021), 1–35.
  37. BiFair: Training Fair Models with Bilevel Optimization. ArXiv preprint abs/2106.04757 (2021). https://arxiv.org/abs/2106.04757
  38. Automatic differentiation in pytorch. (2017).
  39. Benchmarking Bias Mitigation Algorithms in Representation Learning through Fairness Metrics. (2021).
  40. Not All Unlabeled Data are Equal: Learning to Weight Data in Semi-supervised Learning. In Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, NeurIPS 2020, December 6-12, 2020, virtual, Hugo Larochelle, Marc’Aurelio Ranzato, Raia Hadsell, Maria-Florina Balcan, and Hsuan-Tien Lin (Eds.). https://proceedings.neurips.cc/paper/2020/hash/f7ac67a9aa8d255282de7d11391e1b69-Abstract.html
  41. Distributionally Robust Neural Networks. In Proc. of ICLR. OpenReview.net. https://openreview.net/forum?id=ryxGuJrFvS
  42. Fair infinitesimal jackknife: Mitigating the influence of biased training data points without refitting. Advances in Neural Information Processing Systems 35 (2022), 35894–35906.
  43. Shai Shalev-Shwartz and Shai Ben-David. 2014. Understanding machine learning: From theory to algorithms. Cambridge university press.
  44. Interactive Label Cleaning with Example-based Explanations. In Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, NeurIPS 2021, December 6-14, 2021, virtual, Marc’Aurelio Ranzato, Alina Beygelzimer, Yann N. Dauphin, Percy Liang, and Jennifer Wortman Vaughan (Eds.). 12966–12977. https://proceedings.neurips.cc/paper/2021/hash/6c349155b122aa8ad5c877007e05f24f-Abstract.html
  45. Paul Voigt and Axel Von dem Bussche. 2017. The eu general data protection regulation (gdpr). A Practical Guide, 1st Ed., Cham: Springer International Publishing 10, 3152676 (2017), 10–5555.
  46. Avoiding disparate impact with counterfactual distributions. In NeurIPS Workshop on Ethical, Social and Governance Issues in AI.
  47. Understanding Instance-Level Impact of Fairness Constraints. In International Conference on Machine Learning, ICML 2022, 17-23 July 2022, Baltimore, Maryland, USA (Proceedings of Machine Learning Research, Vol. 162), Kamalika Chaudhuri, Stefanie Jegelka, Le Song, Csaba Szepesvári, Gang Niu, and Sivan Sabato (Eds.). PMLR, 23114–23130. https://proceedings.mlr.press/v162/wang22ac.html
  48. Mei Wang and Weihong Deng. 2021. Deep face recognition: A survey. Neurocomputing 429 (2021), 215–244.
  49. Racial Faces in the Wild: Reducing Racial Bias by Information Maximization Adaptation Network. In 2019 IEEE/CVF International Conference on Computer Vision, ICCV 2019, Seoul, Korea (South), October 27 - November 2, 2019. IEEE, 692–702. https://doi.org/10.1109/ICCV.2019.00078
  50. Towards Fairness in Visual Recognition: Effective Strategies for Bias Mitigation. In 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2020, Seattle, WA, USA, June 13-19, 2020. IEEE, 8916–8925. https://doi.org/10.1109/CVPR42600.2020.00894
  51. On the margin theory of feedforward neural networks. (2018).
  52. Ziwei Wu and Jingrui He. 2022. Fairness-aware Model-agnostic Positive and Unlabeled Learning. In 2022 ACM Conference on Fairness, Accountability, and Transparency. 1698–1708.
  53. Fair Class Balancing: Enhancing Model Fairness without Observing Sensitive Attributes. In CIKM ’20: The 29th ACM International Conference on Information and Knowledge Management, Virtual Event, Ireland, October 19-23, 2020, Mathieu d’Aquin, Stefan Dietze, Claudia Hauff, Edward Curry, and Philippe Cudré-Mauroux (Eds.). ACM, 1715–1724. https://doi.org/10.1145/3340531.3411980
  54. Dataset Pruning: Reducing Training Data by Examining Generalization Influence. In The Eleventh International Conference on Learning Representations.
  55. Zhe Yu. 2021. Fair Balance: Mitigating Machine Learning Bias Against Multiple Protected Attributes With Data Balancing. ArXiv preprint abs/2107.08310 (2021). https://arxiv.org/abs/2107.08310
  56. Fairness Beyond Disparate Treatment & Disparate Impact: Learning Classification without Disparate Mistreatment. In Proceedings of the 26th International Conference on World Wide Web, WWW 2017, Perth, Australia, April 3-7, 2017, Rick Barrett, Rick Cummings, Eugene Agichtein, and Evgeniy Gabrilovich (Eds.). ACM, 1171–1180. https://doi.org/10.1145/3038912.3052660
  57. Fairness Constraints: Mechanisms for Fair Classification. In Proceedings of the 20th International Conference on Artificial Intelligence and Statistics, AISTATS 2017, 20-22 April 2017, Fort Lauderdale, FL, USA (Proceedings of Machine Learning Research, Vol. 54), Aarti Singh and Xiaojin (Jerry) Zhu (Eds.). PMLR, 962–970. http://proceedings.mlr.press/v54/zafar17a.html
  58. Mitigating unwanted biases with adversarial learning. In Proceedings of the 2018 AAAI/ACM Conference on AI, Ethics, and Society. 335–340.
  59. Understanding deep learning (still) requires rethinking generalization. Commun. ACM 64, 3 (2021), 107–115.
  60. Understanding Programmatic Weak Supervision via Source-aware Influence Function. ArXiv preprint abs/2205.12879 (2022). https://arxiv.org/abs/2205.12879
  61. Conditional Learning of Fair Representations. In Proc. of ICLR. OpenReview.net. https://openreview.net/forum?id=Hkekl0NFPr
Citations (8)

Summary

We haven't generated a summary for this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Lightbulb On Streamline Icon: https://streamlinehq.com

Continue Learning

We haven't generated follow-up questions for this paper yet.