Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Soft-prompt Tuning for Large Language Models to Evaluate Bias (2306.04735v2)

Published 7 Jun 2023 in cs.CL, cs.AI, and cs.LG

Abstract: Prompting LLMs has gained immense popularity in recent years due to the advantage of producing good results even without the need for labelled data. However, this requires prompt tuning to get optimal prompts that lead to better model performances. In this paper, we explore the use of soft-prompt tuning on sentiment classification task to quantify the biases of LLMs such as Open Pre-trained Transformers (OPT) and Galactica LLM. Since these models are trained on real-world data that could be prone to bias toward certain groups of populations, it is important to identify these underlying issues. Using soft-prompts to evaluate bias gives us the extra advantage of avoiding the human-bias injection that can be caused by manually designed prompts. We check the model biases on different sensitive attributes using the group fairness (bias) and find interesting bias patterns. Since LLMs have been used in the industry in various applications, it is crucial to identify the biases before deploying these models in practice. We open-source our pipeline and encourage industry researchers to adapt our work to their use cases.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Jacob-Junqi Tian (9 papers)
  2. David Emerson (5 papers)
  3. Sevil Zanjani Miyandoab (5 papers)
  4. Deval Pandya (9 papers)
  5. Laleh Seyyed-Kalantari (10 papers)
  6. Faiza Khan Khattak (10 papers)
Citations (6)