Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Perturbation Augmentation for Fairer NLP (2205.12586v2)

Published 25 May 2022 in cs.CL and cs.AI

Abstract: Unwanted and often harmful social biases are becoming ever more salient in NLP research, affecting both models and datasets. In this work, we ask whether training on demographically perturbed data leads to fairer LLMs. We collect a large dataset of human annotated text perturbations and train a neural perturbation model, which we show outperforms heuristic alternatives. We find that (i) LLMs (LMs) pre-trained on demographically perturbed corpora are typically more fair, and (ii) LMs finetuned on perturbed GLUE datasets exhibit less demographic bias on downstream tasks, and (iii) fairness improvements do not come at the expense of performance on downstream tasks. Lastly, we discuss outstanding questions about how best to evaluate the (un)fairness of LLMs. We hope that this exploration of neural demographic perturbation will help drive more improvement towards fairer NLP.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Rebecca Qian (13 papers)
  2. Candace Ross (25 papers)
  3. Jude Fernandes (3 papers)
  4. Eric Smith (28 papers)
  5. Douwe Kiela (85 papers)
  6. Adina Williams (72 papers)
Citations (48)