Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
60 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Safety and Fairness for Content Moderation in Generative Models (2306.06135v1)

Published 9 Jun 2023 in cs.LG and cs.AI

Abstract: With significant advances in generative AI, new technologies are rapidly being deployed with generative components. Generative models are typically trained on large datasets, resulting in model behaviors that can mimic the worst of the content in the training data. Responsible deployment of generative technologies requires content moderation strategies, such as safety input and output filters. Here, we provide a theoretical framework for conceptualizing responsible content moderation of text-to-image generative technologies, including a demonstration of how to empirically measure the constructs we enumerate. We define and distinguish the concepts of safety, fairness, and metric equity, and enumerate example harms that can come in each domain. We then provide a demonstration of how the defined harms can be quantified. We conclude with a summary of how the style of harms quantification we demonstrate enables data-driven content moderation decisions.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Susan Hao (5 papers)
  2. Piyush Kumar (47 papers)
  3. Sarah Laszlo (6 papers)
  4. Shivani Poddar (7 papers)
  5. Bhaktipriya Radharapu (8 papers)
  6. Renee Shelby (12 papers)
Citations (15)