Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Gender Bias in Coreference Resolution (1804.09301v1)

Published 25 Apr 2018 in cs.CL

Abstract: We present an empirical study of gender bias in coreference resolution systems. We first introduce a novel, Winograd schema-style set of minimal pair sentences that differ only by pronoun gender. With these "Winogender schemas," we evaluate and confirm systematic gender bias in three publicly-available coreference resolution systems, and correlate this bias with real-world and textual gender statistics.

Gender Bias in Coreference Resolution

The paper "Gender Bias in Coreference Resolution" presents an empirical investigation into gender bias within coreference resolution systems. Through the development and application of "Winogender schemas," a novel dataset modeled on Winograd schemas, the authors assess gender bias in three distinct coreference resolution systems representing rule-based, statistical, and neural paradigms.

Methodology and Analysis

The authors construct their analysis around a series of sentence pairs differing only by pronoun gender, designed to reveal systematic biases in occupational contexts. Each sentence in their dataset consists of a referring expression related to a person's occupation, alongside another participant and a pronoun, thereby forming a minimal pair schema that evaluates gender bias. The dataset was hand-crafted to include 720 sentences from combinations of 60 occupations, two sentence templates per occupation, variations in participants, and three pronoun genders (female, male, and neutral).

Validation of these schemas was completed using Mechanical Turk, with results indicating strong agreement with intended answers, showcasing the dataset's reliability for detecting gender biases. The schemas specifically spotlight instances where resolutions may hinge on the pronouns' gender, thus revealing gender biases in the systems evaluated.

Results

Evaluations across the rule-based, statistical, and neural systems demonstrate clear patterns of gender bias in coreference resolution. Notably, there was a substantial difference in the systems' treating of pronouns, with a higher tendency to associate male pronouns with occupations. Furthermore, substantial correlation was found between the systems' gender biases and occupational gender statistics from real-world employment data and text.

For example, the rule-based system deviated by 68% in resolving male-female pronoun pairs differently. Statistical and neural models showed similar albeit less pronounced sensitivities. These biases were mirrored in the systems' performance on "gotcha" sentences, which intentionally challenged the systems with occupations against real-world gender statistics, leading to worsened accuracy.

Implications and Future Directions

The finding that coreference systems perpetuate and perhaps magnify existing gender stereotypes as reflected in occupational data has significant implications for future research and development in NLP and AI. It underlines the susceptibility of data-driven models to bias inheritance and amplification, necessitating mitigative strategies.

The paper positions the Winogender schemas as a diagnostic tool to identify gender biases in NLP systems. Yet, it cautions that the absence of detected bias in such tests does not equate to the nonexistence of bias in broader contexts. This highlights the importance of continued research into both bias detection and amelioration strategies, as well as the development of more comprehensive diagnostic tools.

Furthermore, while demonstrating that human judgment can serve as a critical baseline in evaluating AI fairness, the authors acknowledge that human biases should not set the ceiling. Future work could expand beyond occupational biases, incorporating diverse contexts to create a more holistic view of gender bias in language technologies.

The limitations noted by the authors suggest that while efforts to address explicit gender bias are underway, the challenge remains to address implicit biases embedded in data and model architectures. Integration of adversarial training methods or balanced datasets as explored in other studies presents potential paths forward.

Ultimately, this paper contributes valuable insights into the nuanced ways gender bias can manifest in AI systems, serving as a foundation for more equitable NLP technologies. Through meticulous schema development and careful analysis, it illuminates critical aspects of gender bias in coreference resolution systems, paving the way toward more responsible and inclusive AI.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Rachel Rudinger (46 papers)
  2. Jason Naradowsky (19 papers)
  3. Brian Leonard (3 papers)
  4. Benjamin Van Durme (173 papers)
Citations (591)