Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Listen and Speak Fairly: A Study on Semantic Gender Bias in Speech Integrated Large Language Models (2407.06957v1)

Published 9 Jul 2024 in eess.AS, cs.CL, and cs.CY

Abstract: Speech Integrated LLMs (SILLMs) combine LLMs with speech perception to perform diverse tasks, such as emotion recognition to speaker verification, demonstrating universal audio understanding capability. However, these models may amplify biases present in training data, potentially leading to biased access to information for marginalized groups. This work introduces a curated spoken bias evaluation toolkit and corresponding dataset. We evaluate gender bias in SILLMs across four semantic-related tasks: speech-to-text translation (STT), spoken coreference resolution (SCR), spoken sentence continuation (SSC), and spoken question answering (SQA). Our analysis reveals that bias levels are language-dependent and vary with different evaluation methods. Our findings emphasize the necessity of employing multiple approaches to comprehensively assess biases in SILLMs, providing insights for developing fairer SILLM systems.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Yi-Cheng Lin (24 papers)
  2. Tzu-Quan Lin (12 papers)
  3. Chih-Kai Yang (13 papers)
  4. Ke-Han Lu (16 papers)
  5. Wei-Chih Chen (20 papers)
  6. Chun-Yi Kuan (14 papers)
  7. Hung-yi Lee (327 papers)

Summary

We haven't generated a summary for this paper yet.