Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Analyzing the Limits of Self-Supervision in Handling Bias in Language (2112.08637v3)

Published 16 Dec 2021 in cs.CL and cs.AI

Abstract: Prompting inputs with natural language task descriptions has emerged as a popular mechanism to elicit reasonably accurate outputs from large-scale generative LLMs with little to no in-context supervision. This also helps gain insight into how well LLMs capture the semantics of a wide range of downstream tasks purely from self-supervised pre-training on massive corpora of unlabeled text. Such models have naturally also been exposed to a lot of undesirable content like racist and sexist language and there is limited work on awareness of models along these dimensions. In this paper, we define and comprehensively evaluate how well such LLMs capture the semantics of four tasks for bias: diagnosis, identification, extraction and rephrasing. We define three broad classes of task descriptions for these tasks: statement, question, and completion, with numerous lexical variants within each class. We study the efficacy of prompting for each task using these classes and the null task description across several decoding methods and few-shot examples. Our analyses indicate that LLMs are capable of performing these tasks to widely varying degrees across different bias dimensions, such as gender and political affiliation. We believe our work is an important step towards unbiased LLMs by quantifying the limits of current self-supervision objectives at accomplishing such sociologically challenging tasks.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Lisa Bauer (7 papers)
  2. Karthik Gopalakrishnan (34 papers)
  3. Spandana Gella (26 papers)
  4. Yang Liu (2253 papers)
  5. Mohit Bansal (304 papers)
  6. Dilek Hakkani-Tur (94 papers)
Citations (1)

Summary

We haven't generated a summary for this paper yet.