Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
153 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Debiasing Sentence Embedders through Contrastive Word Pairs (2403.18555v1)

Published 27 Mar 2024 in cs.CL

Abstract: Over the last years, various sentence embedders have been an integral part in the success of current machine learning approaches to NLP. Unfortunately, multiple sources have shown that the bias, inherent in the datasets upon which these embedding methods are trained, is learned by them. A variety of different approaches to remove biases in embeddings exists in the literature. Most of these approaches are applicable to word embeddings and in fewer cases to sentence embeddings. It is problematic that most debiasing approaches are directly transferred from word embeddings, therefore these approaches fail to take into account the nonlinear nature of sentence embedders and the embeddings they produce. It has been shown in literature that bias information is still present if sentence embeddings are debiased using such methods. In this contribution, we explore an approach to remove linear and nonlinear bias information for NLP solutions, without impacting downstream performance. We compare our approach to common debiasing methods on classical bias metrics and on bias metrics which take nonlinear information into account.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (22)
  1. Persistent anti-muslim bias in large language models. In Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society, AIES ’21, page 298–306, New York, NY, USA. Association for Computing Machinery.
  2. Man is to computer programmer as woman is to homemaker? debiasing word embeddings. In Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., and Garnett, R., editors, Advances in Neural Information Processing Systems, volume 29, pages 4349–4357. Curran Associates, Inc.
  3. Semantics derived automatically from language corpora contain human-like biases. Science, 356(6334):183–186.
  4. Fairfil: Contrastive neural debiasing method for pretrained text encoders. CoRR, abs/2103.06413.
  5. BERT: pre-training of deep bidirectional transformers for language understanding. Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, abs/1810.04805.
  6. Multi-news: a large-scale multi-document summarization dataset and abstractive hierarchical model.
  7. Lipstick on a pig: Debiasing methods cover up systematic gender biases in word embeddings but do not remove them. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 609–614, Minneapolis, Minnesota. Association for Computational Linguistics.
  8. Towards debiasing sentence representations.
  9. Roberta: A robustly optimized BERT pretraining approach. CoRR, abs/1907.11692.
  10. Black is to criminal as caucasian is to police: Detecting and removing multiclass bias in word embeddings. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1, pages 615–621, Minneapolis, Minnesota. Association for Computational Linguistics.
  11. On measuring social biases in sentence encoders.
  12. Efficient estimation of word representations in vector space.
  13. Learning transferable visual models from natural language supervision. In International Conference on Machine Learning, pages 8748–8763. PMLR.
  14. Hierarchical text-conditional image generation with clip latents. OpenAI papers.
  15. Null it out: Guarding protected attributes by iterative nullspace projection. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 7237–7256, Online. Association for Computational Linguistics.
  16. So can we use intrinsic bias measures or not? In International Conference on Pattern Recognition Applications and Methods.
  17. Deepview: Visualizing classification boundaries of deep neural networks as scatter plots using discriminative dimensionality reduction. Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence.
  18. Attention is all you need. In Guyon, I., Luxburg, U. V., Bengio, S., Wallach, H., Fergus, R., Vishwanathan, S., and Garnett, R., editors, Advances in Neural Information Processing Systems, volume 30. Curran Associates, Inc.
  19. GLUE: A multi-task benchmark and analysis platform for natural language understanding. In the Proceedings of ICLR.
  20. Huggingface’s transformers: State-of-the-art natural language processing.
  21. Gender bias in contextualized word embeddings.
  22. Gender bias in coreference resolution: Evaluation and debiasing methods. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 2, pages 15–20, New Orleans, Louisiana. Association for Computational Linguistics.

Summary

We haven't generated a summary for this paper yet.

Youtube Logo Streamline Icon: https://streamlinehq.com