Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Stereotype and Skew: Quantifying Gender Bias in Pre-trained and Fine-tuned Language Models (2101.09688v2)

Published 24 Jan 2021 in cs.CL, cs.AI, cs.LG, and cs.NE

Abstract: This paper proposes two intuitive metrics, skew and stereotype, that quantify and analyse the gender bias present in contextual LLMs when tackling the WinoBias pronoun resolution task. We find evidence that gender stereotype correlates approximately negatively with gender skew in out-of-the-box models, suggesting that there is a trade-off between these two forms of bias. We investigate two methods to mitigate bias. The first approach is an online method which is effective at removing skew at the expense of stereotype. The second, inspired by previous work on ELMo, involves the fine-tuning of BERT using an augmented gender-balanced dataset. We show that this reduces both skew and stereotype relative to its unaugmented fine-tuned counterpart. However, we find that existing gender bias benchmarks do not fully probe professional bias as pronoun resolution may be obfuscated by cross-correlations from other manifestations of gender prejudice. Our code is available online, at https://github.com/12kleingordon34/NLP_masters_project.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Daniel de Vassimon Manela (4 papers)
  2. David Errington (5 papers)
  3. Thomas Fisher (2 papers)
  4. Boris van Breugel (18 papers)
  5. Pasquale Minervini (88 papers)
Citations (78)