Leveraging Diffusion Perturbations for Measuring Fairness in Computer Vision (2311.15108v2)
Abstract: Computer vision models have been known to encode harmful biases, leading to the potentially unfair treatment of historically marginalized groups, such as people of color. However, there remains a lack of datasets balanced along demographic traits that can be used to evaluate the downstream fairness of these models. In this work, we demonstrate that diffusion models can be leveraged to create such a dataset. We first use a diffusion model to generate a large set of images depicting various occupations. Subsequently, each image is edited using inpainting to generate multiple variants, where each variant refers to a different perceived race. Using this dataset, we benchmark several vision-LLMs on a multi-class occupation classification task. We find that images generated with non-Caucasian labels have a significantly higher occupation misclassification rate than images generated with Caucasian labels, and that several misclassifications are suggestive of racial biases. We measure a model's downstream fairness by computing the standard deviation in the probability of predicting the true occupation label across the different perceived identity groups. Using this fairness metric, we find significant disparities between the evaluated vision-and-LLMs. We hope that our work demonstrates the potential value of diffusion methods for fairness evaluations.
- Turning a blind eye: Explicit removal of biases and variation from deep neural network embeddings. In Proceedings of the European Conference on Computer Vision Workshops.
- The Problem with Bias: From Allocative to Representational Harms in Machine Learning. In Special Interest Group for Computing, Information and Society (SIGCIS).
- Towards Language Models That Can See: Computer Vision Through the LENS of Natural Language. arXiv:2306.16410.
- Multimodal datasets: misogyny, pornography, and malignant stereotypes. arXiv preprint arXiv:2110.01963.
- Gender Shades: Intersectional Accuracy Disparities in Commercial Gender Classification. In Proceedings of Machine Learning Research.
- Domain balancing: Face recognition on long-tailed domains. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition.
- Evaluating and mitigating bias in image classifiers: A causal perspective using counterfactuals. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision.
- Diffusion models beat gans on image synthesis. In Proceedings of Advances in Neural Information Processing Systems,.
- Cyberbullying classifiers are sensitive to model-agnostic perturbations. arXiv preprint arXiv:2201.06384.
- Fair diffusion: Instructing text-to-image generation models on fairness. arXiv preprint arXiv:2302.10893.
- Glover, K. S. 2009. Racial profiling: Research, racism, and resistance. Rowman & Littlefield Publishers.
- Making the v in vqa matter: Elevating the role of image understanding in visual question answering. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.
- Measuring individual differences in implicit cognition: the implicit association test. Journal of personality and social psychology.
- TIFA: Accurate and interpretable text-to-image faithfulness evaluation with question answering. arXiv preprint arXiv:2303.11897.
- Zero-shot racially balanced dataset generation using an existing biased StyleGAN2. arXiv preprint arXiv:2305.07710.
- Fairface: Face attribute dataset for balanced race, gender, and age for bias measurement and mitigation. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision.
- ViLT: Vision-and-language transformer without convolution or region supervision. In International Conference on Machine Learning.
- Segment anything. arXiv preprint arXiv:2304.02643.
- Grounding DINO: Marrying DINO with grounded pre-training for open-set object detection. arXiv preprint arXiv:2303.05499.
- Stable bias: Analyzing societal representations in diffusion models. arXiv preprint arXiv:2303.11408.
- Dynaboard: An evaluation-as-a-service platform for holistic next-generation benchmarking. In Proceedings of Advances in Neural Information Processing Systems.
- Racial appearance bias: Improving evidence-based policies to address racial disparities. Policy Insights from the Behavioral and Brain Sciences.
- A decade of studying implicit racial/ethnic bias in healthcare providers using the implicit association test. Social science & medicine.
- It’s all in the name: Mitigating gender bias with name-based counterfactual data substitution. In Proceedings of the Conference on Empirical Methods in Natural Language Processing and the International Joint Conference on Natural Language Processing (EMNLP-IJCNLP).
- SDEdit: Guided image synthesis and editing with stochastic differential equations. In International Conference on Learning Representations.
- SDXL: Improving Latent Diffusion Models for High-Resolution Image Synthesis. arXiv preprint arXiv:2307.01952.
- Perturbation augmentation for fairer nlp. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing.
- Learning transferable visual models from natural language supervision. In International Conference on Machine Learning.
- High-resolution image synthesis with latent diffusion models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition.
- FLAVA: A foundational language and vision alignment model. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition.
- Hi, my name is Martha: Using names to measure and mitigate bias in generative dialogue models. arXiv preprint arXiv:2109.03300.
- Deep unsupervised learning using nonequilibrium thermodynamics. In International Conference on Machine Learning.
- Convnets and imagenet beyond accuracy: Understanding mistakes and uncovering biases. In Proceedings of the European Conference on Computer Vision.
- Dynatask: A framework for creating dynamic AI benchmark tasks. In Proceedings of the Association for Computational Linguistics: System Demonstrations.
- Mitigating bias in face recognition using skewness-aware reinforcement learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition.
- Imagen editor and editbench: Advancing and evaluating text-guided image inpainting. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition.
- Balanced Datasets Are Not Enough: Estimating and Mitigating Gender Bias in Deep Image Representations. In Proceedings of the IEEE/CVF International Conference on Computer Vision.
- Exploring racial bias within face recognition via per-subject adversarially-enabled data augmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops.
- Age Progression/Regression by Conditional Adversarial Autoencoder. In IEEE Conference on Computer Vision and Pattern Recognition.