Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

UnQovering Stereotyping Biases via Underspecified Questions (2010.02428v3)

Published 6 Oct 2020 in cs.CL

Abstract: While language embeddings have been shown to have stereotyping biases, how these biases affect downstream question answering (QA) models remains unexplored. We present UNQOVER, a general framework to probe and quantify biases through underspecified questions. We show that a naive use of model scores can lead to incorrect bias estimates due to two forms of reasoning errors: positional dependence and question independence. We design a formalism that isolates the aforementioned errors. As case studies, we use this metric to analyze four important classes of stereotypes: gender, nationality, ethnicity, and religion. We probe five transformer-based QA models trained on two QA datasets, along with their underlying LLMs. Our broad study reveals that (1) all these models, with and without fine-tuning, have notable stereotyping biases in these classes; (2) larger models often have higher bias; and (3) the effect of fine-tuning on bias varies strongly with the dataset and the model size.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Tao Li (441 papers)
  2. Tushar Khot (53 papers)
  3. Daniel Khashabi (83 papers)
  4. Ashish Sabharwal (84 papers)
  5. Vivek Srikumar (55 papers)
Citations (115)

Summary

We haven't generated a summary for this paper yet.