Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Walking in Others' Shoes: How Perspective-Taking Guides Large Language Models in Reducing Toxicity and Bias (2407.15366v1)

Published 22 Jul 2024 in cs.CL, cs.AI, and cs.CY

Abstract: The common toxicity and societal bias in contents generated by LLMs necessitate strategies to reduce harm. Present solutions often demand white-box access to the model or substantial training, which is impractical for cutting-edge commercial LLMs. Moreover, prevailing prompting methods depend on external tool feedback and fail to simultaneously lessen toxicity and bias. Motivated by social psychology principles, we propose a novel strategy named \textbf{perspective-taking prompting (\textsc{PeT})} that inspires LLMs to integrate diverse human perspectives and self-regulate their responses. This self-correction mechanism can significantly diminish toxicity (up to $89\%$) and bias (up to $73\%$) in LLMs' responses. Rigorous evaluations and ablation studies are conducted on two commercial LLMs (ChatGPT and GLM) and three open-source LLMs, revealing \textsc{PeT}'s superiority in producing less harmful responses, outperforming five strong baselines.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Rongwu Xu (19 papers)
  2. Zi'an Zhou (1 paper)
  3. Tianwei Zhang (199 papers)
  4. Zehan Qi (13 papers)
  5. Su Yao (34 papers)
  6. Ke Xu (309 papers)
  7. Wei Xu (535 papers)
  8. Han Qiu (60 papers)
Citations (3)