Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash 99 tok/s
Gemini 2.5 Pro 55 tok/s Pro
GPT-5 Medium 23 tok/s
GPT-5 High 19 tok/s Pro
GPT-4o 108 tok/s
GPT OSS 120B 465 tok/s Pro
Kimi K2 179 tok/s Pro
2000 character limit reached

Bias Mitigation in Fine-tuning Pre-trained Models for Enhanced Fairness and Efficiency (2403.00625v1)

Published 1 Mar 2024 in cs.LG and cs.CY

Abstract: Fine-tuning pre-trained models is a widely employed technique in numerous real-world applications. However, fine-tuning these models on new tasks can lead to unfair outcomes. This is due to the absence of generalization guarantees for fairness properties, regardless of whether the original pre-trained model was developed with fairness considerations. To tackle this issue, we introduce an efficient and robust fine-tuning framework specifically designed to mitigate biases in new tasks. Our empirical analysis shows that the parameters in the pre-trained model that affect predictions for different demographic groups are different, so based on this observation, we employ a transfer learning strategy that neutralizes the importance of these influential weights, determined using Fisher information across demographic groups. Additionally, we integrate this weight importance neutralization strategy with a matrix factorization technique, which provides a low-rank approximation of the weight matrix using fewer parameters, reducing the computational demands. Experiments on multiple pre-trained models and new tasks demonstrate the effectiveness of our method.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (38)
  1. Online embedding compression for text classification using low rank matrix factorization, 2018.
  2. A reductions approach to fair classification. CoRR, abs/1803.02453, 2018.
  3. Intrinsic dimensionality explains the effectiveness of language model fine-tuning. In Chengqing Zong, Fei Xia, Wenjie Li, and Roberto Navigli, editors, Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 7319–7328, Online, August 2021. Association for Computational Linguistics.
  4. Compressing pre-trained language models by matrix decomposition. In Kam-Fai Wong, Kevin Knight, and Hua Wu, editors, Proceedings of the 1st Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics and the 10th International Joint Conference on Natural Language Processing, pages 884–889, Suzhou, China, December 2020. Association for Computational Linguistics.
  5. Fairness Constraints: Mechanisms for Fair Classification. arXiv e-prints, page arXiv:1507.05259, Jul 2015.
  6. Fairness Beyond Disparate Treatment & Disparate Impact: Learning Classification without Disparate Mistreatment. arXiv e-prints, page arXiv:1610.08452, Oct 2016.
  7. Exploring or exploiting? social and ethical implications of autonomous experimentation in ai. In Workshop on Fairness, Accountability, and Transparency in Machine Learning (FAT-ML), New York University, page 4, October 2016.
  8. Gender shades: Intersectional accuracy disparities in commercial gender classification. In Sorelle A. Friedler and Christo Wilson, editors, Proceedings of the 1st Conference on Fairness, Accountability and Transparency, volume 81 of Proceedings of Machine Learning Research, pages 77–91, New York, NY, USA, 23–24 Feb 2018. PMLR.
  9. Optimized pre-processing for discrimination prevention. In I. Guyon, U. V. Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett, editors, Advances in Neural Information Processing Systems 30, pages 3992–4001. Curran Associates, Inc., 2017.
  10. Flexibly fair representation learning by disentanglement. CoRR, abs/1906.02589, 2019.
  11. The accuracy, fairness, and limits of predicting recidivism. Science Advances, 4(1):eaao5580, 2018.
  12. Fairness Through Awareness. arXiv e-prints, page arXiv:1104.3913, Apr 2011.
  13. On the effectiveness of parameter-efficient fine-tuning. In Proceedings of the Thirty-Seventh AAAI Conference on Artificial Intelligence and Thirty-Fifth Conference on Innovative Applications of Artificial Intelligence and Thirteenth Symposium on Educational Advances in Artificial Intelligence, AAAI’23/IAAI’23/EAAI’23. AAAI Press, 2023.
  14. Singular value decomposition and least squares solutions. Linear Algebra, pages 134–151, 1971.
  15. Equality of opportunity in supervised learning. CoRR, abs/1610.02413, 2016.
  16. Deep residual learning for image recognition. In 2016 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2016, Las Vegas, NV, USA, June 27-30, 2016, pages 770–778. IEEE Computer Society, 2016.
  17. Towards a unified view of parameter-efficient transfer learning. In International Conference on Learning Representations, 2022.
  18. Language model compression with weighted low-rank factorization. In International Conference on Learning Representations, 2022.
  19. LoRA: Low-rank adaptation of large language models. In International Conference on Learning Representations, 2022.
  20. Densely connected convolutional networks. In CVPR, pages 2261–2269. IEEE Computer Society, 2017.
  21. Adafair: Cumulative fairness adaptive boosting. In Proceedings of the 28th ACM International Conference on Information and Knowledge Management, CIKM ’19, page 781–790, New York, NY, USA, 2019. Association for Computing Machinery.
  22. Speeding up convolutional neural networks with low rank expansions, 2014.
  23. Fairness-aware classifier with prejudice remover regularizer. In Peter A. Flach, Tijl De Bie, and Nello Cristianini, editors, Machine Learning and Knowledge Discovery in Databases, pages 35–50, Berlin, Heidelberg, 2012. Springer Berlin Heidelberg.
  24. Consumer credit-risk models via machine-learning algorithms. Journal of Banking & Finance, 34(11):2767 – 2787, 2010.
  25. Sex- and gender-specific disparities in colorectal cancer risk. World journal of gastroenterology : WJG, 21:5167–5175, 05 2015.
  26. Ron Kohavi. Scaling up the accuracy of naive-bayes classifiers: A decision-tree hybrid. In Proceedings of the Second International Conference on Knowledge Discovery and Data Mining, KDD’96, page 202–207. AAAI Press, 1996.
  27. Exploring versatile generative language model via parameter-efficient transfer learning. In Trevor Cohn, Yulan He, and Yang Liu, editors, Findings of the Association for Computational Linguistics: EMNLP 2020, pages 441–459, Online, November 2020. Association for Computational Linguistics.
  28. Deep learning face attributes in the wild. In 2015 IEEE International Conference on Computer Vision (ICCV), pages 3730–3738, 2015.
  29. Roberta: A robustly optimized bert pretraining approach, 2019. cite arxiv:1907.11692.
  30. Exploiting mmd and sinkhorn divergences for fair and transferable representation learning. In H. Larochelle, M. Ranzato, R. Hadsell, M.F. Balcan, and H. Lin, editors, Advances in Neural Information Processing Systems, volume 33, pages 15360–15370. Curran Associates, Inc., 2020.
  31. Revisiting natural gradient for deep networks. In Yoshua Bengio and Yann LeCun, editors, 2nd International Conference on Learning Representations, ICLR 2014, Banff, AB, Canada, April 14-16, 2014, Conference Track Proceedings, 2014.
  32. FR-train: A mutual information-based approach to fair and robust training. In Hal Daumé III and Aarti Singh, editors, Proceedings of the 37th International Conference on Machine Learning, volume 119 of Proceedings of Machine Learning Research, pages 8147–8157. PMLR, 13–18 Jul 2020.
  33. Towards fine-tuning pre-trained language models with integer forward and backward propagation. In Andreas Vlachos and Isabelle Augenstein, editors, Findings of the Association for Computational Linguistics: EACL 2023, pages 1912–1921, Dubrovnik, Croatia, May 2023. Association for Computational Linguistics.
  34. Effective Unconstrained Face Recognition by Combining Multiple Descriptors and Learned Background Statistics. IEEE Transactions on Pattern Analysis and Machine Intelligence, 33(10):1978–1990, 2011.
  35. Raise a child in large language model: Towards effective and generalizable fine-tuning. In Marie-Francine Moens, Xuanjing Huang, Lucia Specia, and Scott Wen-tau Yih, editors, Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 9514–9528, Online and Punta Cana, Dominican Republic, November 2021. Association for Computational Linguistics.
  36. Learning fair representations. In Sanjoy Dasgupta and David McAllester, editors, Proceedings of the 30th International Conference on Machine Learning, volume 28 of Proceedings of Machine Learning Research, pages 325–333, Atlanta, Georgia, USA, 17–19 Jun 2013. PMLR.
  37. Mitigating unwanted biases with adversarial learning. In Proceedings of the 2018 AAAI/ACM Conference on AI, Ethics, and Society, AIES ’18, page 335–340, New York, NY, USA, 2018. Association for Computing Machinery.
  38. Fair representation learning with unreliable labels. In Francisco Ruiz, Jennifer Dy, and Jan-Willem van de Meent, editors, Proceedings of The 26th International Conference on Artificial Intelligence and Statistics, volume 206 of Proceedings of Machine Learning Research, pages 4655–4667. PMLR, 25–27 Apr 2023.
Citations (2)
List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Summary

We haven't generated a summary for this paper yet.

Ai Generate Text Spark Streamline Icon: https://streamlinehq.com

Paper Prompts

Sign up for free to create and run prompts on this paper using GPT-5.

Dice Question Streamline Icon: https://streamlinehq.com

Follow-up Questions

We haven't generated follow-up questions for this paper yet.

Authors (2)

Don't miss out on important new AI/ML research

See which papers are being discussed right now on X, Reddit, and more:

“Emergent Mind helps me see which AI papers have caught fire online.”

Philip

Philip

Creator, AI Explained on YouTube