Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Language Generation Models Can Cause Harm: So What Can We Do About It? An Actionable Survey (2210.07700v2)

Published 14 Oct 2022 in cs.CL

Abstract: Recent advances in the capacity of LLMs to generate human-like text have resulted in their increased adoption in user-facing settings. In parallel, these improvements have prompted a heated discourse around the risks of societal harms they introduce, whether inadvertent or malicious. Several studies have explored these harms and called for their mitigation via development of safer, fairer models. Going beyond enumerating the risks of harms, this work provides a survey of practical methods for addressing potential threats and societal harms from language generation models. We draw on several prior works' taxonomies of LLM risks to present a structured overview of strategies for detecting and ameliorating different kinds of risks/harms of language generators. Bridging diverse strands of research, this survey aims to serve as a practical guide for both LM researchers and practitioners, with explanations of different mitigation strategies' motivations, their limitations, and open problems for future research.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Sachin Kumar (68 papers)
  2. Vidhisha Balachandran (31 papers)
  3. Lucille Njoo (3 papers)
  4. Antonios Anastasopoulos (111 papers)
  5. Yulia Tsvetkov (142 papers)
Citations (70)