Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Language-Agnostic Bias Detection in Language Models with Bias Probing (2305.13302v2)

Published 22 May 2023 in cs.CL

Abstract: Pretrained LLMs (PLMs) are key components in NLP, but they contain strong social biases. Quantifying these biases is challenging because current methods focusing on fill-the-mask objectives are sensitive to slight changes in input. To address this, we propose a bias probing technique called LABDet, for evaluating social bias in PLMs with a robust and language-agnostic method. For nationality as a case study, we show that LABDet `surfaces' nationality bias by training a classifier on top of a frozen PLM on non-nationality sentiment detection. We find consistent patterns of nationality bias across monolingual PLMs in six languages that align with historical and political context. We also show for English BERT that bias surfaced by LABDet correlates well with bias in the pretraining data; thus, our work is one of the few studies that directly links pretraining data to PLM behavior. Finally, we verify LABDet's reliability and applicability to different templates and languages through an extensive set of robustness checks. We publicly share our code and dataset in https://github.com/akoksal/LABDet.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Abdullatif Köksal (22 papers)
  2. Omer Faruk Yalcin (1 paper)
  3. Ahmet Akbiyik (1 paper)
  4. M. Tahir Kilavuz (1 paper)
  5. Anna Korhonen (90 papers)
  6. Hinrich Schütze (250 papers)
Citations (1)

Summary

We haven't generated a summary for this paper yet.