Revealing and Reducing Gender Biases in Vision and Language Assistants (VLAs) (2410.19314v2)
Abstract: Pre-trained LLMs have been reliably integrated with visual input for multimodal tasks. The widespread adoption of instruction-tuned image-to-text vision-language assistants (VLAs) like LLaVA and InternVL necessitates evaluating gender biases. We study gender bias in 22 popular open-source VLAs with respect to personality traits, skills, and occupations. Our results show that VLAs replicate human biases likely present in the data, such as real-world occupational imbalances. Similarly, they tend to attribute more skills and positive personality traits to women than to men, and we see a consistent tendency to associate negative personality traits with men. To eliminate the gender bias in these models, we find that fine-tuning-based debiasing methods achieve the best trade-off between debiasing and retaining performance on downstream tasks. We argue for pre-deploying gender bias assessment in VLAs and motivate further development of debiasing strategies to ensure equitable societal outcomes. Code is available at https://github.com/ExplainableML/vla-gender-bias.
- Phi-3 technical report: A highly capable language model locally on your phone. In arXiv, 2024.
- Gpt-4 technical report. In arXiv, 2023.
- Evaluating clip: towards characterization of broader capabilities and downstream implications. In arXiv, 2021.
- Flamingo: a visual language model for few-shot learning. In NeurIPS, 2022.
- Qwen-vl: A versatile vision-language model for understanding, localization, text reading, and beyond. In arXiv, 2023.
- Intersectionality in quantitative research: A systematic review of its emergence and applications of theory and methods. In SSM-population health, 2021.
- On the dangers of stochastic parrots: Can language models be too big? In ACM conference on fairness, accountability, and transparency, 2021.
- A prompt array keeps the bias away: Debiasing vision-language models with adversarial learning. In AACL-IJCNLP, 2022.
- Easily accessible text-to-image generation amplifies demographic stereotypes at large scale. In ACM Conference on Fairness, Accountability, and Transparency, 2023.
- Internlm2 technical report. In arXiv, 2024.
- How far are we to gpt-4v? closing the gap to commercial multimodal models with open-source suites. In arXiv, 2024a.
- Internvl: Scaling up vision foundation models and aligning for generic visual-linguistic tasks. In CVPR, 2024b.
- Reproducible scaling laws for contrastive language-image learning. In CVPR, 2023.
- Vicuna: An open-source chatbot impressing gpt-4 with 90%* chatgpt quality, 2023.
- Mobilevlm v2: Faster and stronger baseline for vision language model. In arXiv, 2024.
- Vlmevalkit: An open-source toolkit for evaluating large multi-modality models. In arXiv, 2024.
- All our n-gram are belong to you, 2006.
- Examining gender and racial bias in large vision–language models using a novel dataset of parallel images. In EACL, 2024.
- Mme: A comprehensive evaluation benchmark for multimodal large language models. In arXiv, 2023.
- Bias and fairness in large language models: A survey. In Computational Linguistics, 2024.
- Uncurated image-text datasets: Shedding light on demographic bias. In CVPR, 2023.
- Gemini: a family of highly capable multimodal models. In arXiv, 2023.
- Unboxing occupational bias: Grounded debiasing llms with us labor data. In arXiv, 2024.
- Visogender: A dataset for benchmarking gender bias in image-text pronoun resolution. In NeurIPS, 2024.
- Efficient multimodal learning from data-centric perspective. In arXiv, 2024.
- Madeline E Heilman. Gender stereotypes and workplace bias. In Research in organizational Behavior, 2012.
- Women at work: pathways from gender stereotypes to gender bias and discrimination. In Annual Review of Organizational Psychology and Organizational Behavior, 2024.
- Social bias evaluation for large language models requires prompt variations. In arXiv, 2024.
- Quantifying societal bias amplification in image captioning. In CVPR, 2022.
- Uncovering bias in large vision-language models at scale with counterfactuals. In arXiv, 2024.
- Unsupervised discovery of gendered language through latent-variable modeling. In ACL, 2019.
- LoRA: Low-rank adaptation of large language models. In ICLR, 2022.
- Multi-modal bias: Introducing a framework for stereotypical bias assessment beyond gender and race in vision–language models. In EACL, 2023.
- Kameni Florentin Flambeau Jiechieu and Norbert Tsopze. Skills prediction based on multi-label resume classification using cnn with model predictions explanation. In Neural Computing and Applications, 2021.
- Fairface: Face attribute dataset for balanced race, gender, and age for bias measurement and mitigation. In WACV, 2021.
- Measuring bias in contextualized word representations. In First Workshop on Gender Bias in Natural Language Processing, 2019.
- Optimal brain damage. In NeurIPS, 1989.
- The power of scale for parameter-efficient prompt tuning. In EMNLP, 2021.
- Seed-bench-2: Benchmarking multimodal large language models. In arXiv, 2023a.
- Seed-bench: Benchmarking multimodal large language models. In CVPR, 2024a.
- Blip-2: Bootstrapping language-image pre-training with frozen image encoders and large language models. In ICML, 2023b.
- Red teaming visual language models. In Findings of the Association for Computational Linguistics, 2024b.
- Improved baselines with visual instruction tuning. In CVPR, 2024a.
- Llava-next: Improved reasoning, ocr, and world knowledge, 2024b.
- Visual instruction tuning. In NeurIPS, 2024c.
- Mmbench: Is your multi-modal model an all-around player? In arXiv, 2023.
- Stable bias: Evaluating societal representations in diffusion models. In NeurIPS, 2024.
- Llm-pruner: On the structural pruning of large language models. In NeurIPS, 2023.
- Debiasing surgeon: fantastic weights and how to find them. In arXiv, 2024.
- Moses Olafenwa. Idenprof, 2018.
- Learning transferable visual models from natural language supervision. In ICML, 2021.
- Status incongruity and backlash effects: Defending the gender hierarchy motivates prejudice against female leaders. In Journal of experimental social psychology, 2012.
- A multi-dimensional study on bias in vision-language models. In Findings of the Association for Computational Linguistics, 2023.
- A unified framework and dataset for assessing gender bias in vision-language models. In arXiv, 2024.
- A step toward more inclusive people annotations for fairness. In AIES, 2021.
- Quantifying language models’ sensitivity to spurious features in prompt design or: How i learned to start worrying about prompt formatting. In ICLR, 2024.
- Quantifying social biases using templates is unreliable. In NeurIPS Workshop on Trustworthy and Socially Responsible Machine Learning, 2022.
- The bias amplification paradox in text-to-image generation. In NAACL, 2024.
- Dear: Debiasing vision-language models with additive residuals. In CVPR, 2023.
- SkunkworksAI. Bakllava, 2023.
- Aligning large multimodal models with factually augmented rlhf. In Findings of the Association for Computational Linguistics, 2024.
- Discovering and mitigating biases in clip-based image editing. In WACV, 2024.
- Cambrian-1: A fully open, vision-centric exploration of multimodal llms. In arXiv, 2024.
- Stable diffusion exposed: Gender bias from prompt to image. In arXiv, 2023.
- Genderbias-vl: Benchmarking gender bias in vision language models via counterfactual probing. In arXiv, 2024.
- Gendered skills and unemployed men’s resistance to “women’s work”. In Gender, Work & Organization, 2021.
- Mmmu: A massive multi-discipline multimodal understanding and reasoning benchmark for expert agi. In CVPR, 2024.
- Sigmoid loss for language image pre-training. In ICCV, 2023.
- Vlbiasbench: A comprehensive benchmark for evaluating bias in large vision-language model. In arXiv, 2024.
- Gender bias in coreference resolution: Evaluation and debiasing methods. In NAACL, 2018.
- Vlstereoset: A study of stereotypical bias in pre-trained vision-language models. In AACL-IJCNLP, 2022.
- Minigpt-4: Enhancing vision-language understanding with advanced large language models. In ICLR, 2024.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.