Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
51 tokens/sec
GPT-4o
60 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Uncovering Bias in Large Vision-Language Models with Counterfactuals (2404.00166v2)

Published 29 Mar 2024 in cs.CV and cs.AI

Abstract: With the advent of LLMs possessing increasingly impressive capabilities, a number of Large Vision-LLMs (LVLMs) have been proposed to augment LLMs with visual inputs. Such models condition generated text on both an input image and a text prompt, enabling a variety of use cases such as visual question answering and multimodal chat. While prior studies have examined the social biases contained in text generated by LLMs, this topic has been relatively unexplored in LVLMs. Examining social biases in LVLMs is particularly challenging due to the confounding contributions of bias induced by information contained across the text and visual modalities. To address this challenging problem, we conduct a large-scale study of text generated by different LVLMs under counterfactual changes to input images. Specifically, we present LVLMs with identical open-ended text prompts while conditioning on images from different counterfactual sets, where each set contains images which are largely identical in their depiction of a common subject (e.g., a doctor), but vary only in terms of intersectional social attributes (e.g., race and gender). We comprehensively evaluate the text produced by different LVLMs under this counterfactual generation setting and find that social attributes such as race, gender, and physical characteristics depicted in input images can significantly influence toxicity and the generation of competency-associated words.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (15)
  1. Gpt-4 technical report. arXiv preprint arXiv:2303.08774, 2023.
  2. Toward gender-inclusive coreference resolution. arXiv preprint arXiv:1910.13913, 2019.
  3. Susan T Fiske. Stereotype content: Warmth and competence endure. Current directions in psychological science, 27(2):67–73, 2018.
  4. Examining gender and racial bias in large vision-language models using a novel dataset of parallel images. arXiv preprint arXiv:2402.05779, 2024.
  5. Probing and mitigating intersectional social biases in vision-language models with counterfactual examples. arXiv preprint arXiv:2312.00825, 2023.
  6. Visual instruction tuning. Advances in neural information processing systems, 36, 2024.
  7. Bias against 93 stigmatized groups in masked language models and downstream sentiment classification tasks. In Proceedings of the 2023 ACM Conference on Fairness, Accountability, and Transparency, pages 1699–1710, 2023.
  8. Stereoset: Measuring stereotypical bias in pretrained language models. arXiv preprint arXiv:2004.09456, 2020.
  9. Crows-pairs: A challenge dataset for measuring social biases in masked language models. arXiv preprint arXiv:2010.00133, 2020.
  10. Comprehensive stereotype content dictionaries using a semi-automated method. European Journal of Social Psychology, 51(1):178–196, 2021.
  11. Learning transferable visual models from natural language supervision. In International conference on machine learning, pages 8748–8763. PMLR, 2021.
  12. Gender bias in coreference resolution. arXiv preprint arXiv:1804.09301, 2018.
  13. “i’m sorry to hear that”: Finding new biases in language models with a holistic descriptor dataset. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pages 9180–9211, 2022.
  14. Mind the gap: A balanced corpus of gendered ambiguous pronouns. Transactions of the Association for Computational Linguistics, 6:605–617, 2018.
  15. Gender bias in coreference resolution: Evaluation and debiasing methods. arXiv preprint arXiv:1804.06876, 2018.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Phillip Howard (28 papers)
  2. Anahita Bhiwandiwalla (15 papers)
  3. Kathleen C. Fraser (22 papers)
  4. Svetlana Kiritchenko (34 papers)
Citations (4)
X Twitter Logo Streamline Icon: https://streamlinehq.com