Survey of Social Bias in Vision-Language Models (2309.14381v1)
Abstract: In recent years, the rapid advancement of ML models, particularly transformer-based pre-trained models, has revolutionized NLP and Computer Vision (CV) fields. However, researchers have discovered that these models can inadvertently capture and reinforce social biases present in their training datasets, leading to potential social harms, such as uneven resource allocation and unfair representation of specific social groups. Addressing these biases and ensuring fairness in AI systems has become a critical concern in the ML community. The recent introduction of pre-trained vision-and-language (VL) models in the emerging multimodal field demands attention to the potential social biases present in these models as well. Although VL models are susceptible to social bias, there is a limited understanding compared to the extensive discussions on bias in NLP and CV. This survey aims to provide researchers with a high-level insight into the similarities and differences of social bias studies in pre-trained models across NLP, CV, and VL. By examining these perspectives, the survey aims to offer valuable guidelines on how to approach and mitigate social bias in both unimodal and multimodal settings. The findings and recommendations presented here can benefit the ML community, fostering the development of fairer and non-biased AI models in various applications and research endeavors.
- [n. d.]. https://www.washington.edu/news/2015/04/09/whos-a-ceo-google-image-results-can-shift-gender-biases/
- 2015. https://www.bbc.com/news/technology-33347866
- 2019. https://www.oecd.org/digital/artificial-intelligence/
- 2019. https://digital-strategy.ec.europa.eu/en/library/ethics-guidelines-trustworthy-ai
- 2023. https://en.wikipedia.org/wiki/Protected_group
- 2023. https://en.wikipedia.org/wiki/Tay_(chatbot)
- Counterfactual vision and language learning. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 10044–10054.
- Hervé Abdi and Lynne J. Williams. 2010. Principal component analysis. WIREs Computational Statistics 2, 4 (2010), 433–459. https://doi.org/10.1002/wics.101 arXiv:https://wires.onlinelibrary.wiley.com/doi/pdf/10.1002/wics.101
- Evaluating clip: towards characterization of broader capabilities and downstream implications. arXiv preprint arXiv:2108.02818 abs/2108.02818 (2021).
- On Measuring Social Biases in Prompt-Based Multi-Task Learning. In Findings of the Association for Computational Linguistics: NAACL 2022. Association for Computational Linguistics, Seattle, United States, 551–564. https://doi.org/10.18653/v1/2022.findings-naacl.42
- Turning a Blind Eye: Explicit Removal of Biases and Variation from Deep Neural Network Embeddings. In Computer Vision – ECCV 2018 Workshops: Munich, Germany, September 8-14, 2018, Proceedings, Part I (Munich, Germany). Springer-Verlag, Berlin, Heidelberg, 556–572. https://doi.org/10.1007/978-3-030-11009-3_34
- Convolutional Image Captioning. In 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). IEEE Computer Society, Los Alamitos, CA, USA, 5561–5570. https://doi.org/10.1109/CVPR.2018.00583
- Rényi Fair Inference. In International Conference on Learning Representations. https://openreview.net/forum?id=HkgsUJrtDB
- Assessing Political Prudence of Open-domain Chatbots. In Proceedings of the 22nd Annual Meeting of the Special Interest Group on Discourse and Dialogue. Association for Computational Linguistics, Singapore and Online, 548–555. https://aclanthology.org/2021.sigdial-1.57
- The problem with bias: Allocative versus representational harms in machine learning. In 9th Annual conference of the special interest group for computing, information and society.
- Relational inductive biases, deep learning, and graph networks. arXiv (2018). https://arxiv.org/pdf/1806.01261.pdf
- A prompt array keeps the bias away: Debiasing vision-language models with adversarial learning. arXiv preprint arXiv:2203.11933 (2022).
- Mechanisms of linguistic bias: How words reflect and maintain stereotypic expectancies. Social cognition and communication 31 (2014), 313–330.
- Shruti Bhargava and David Forsyth. 2019. Exposing and correcting the gender bias in image captioning datasets and models. arXiv preprint arXiv:1912.00578 (2019).
- Reuben Binns. 2018. Fairness in machine learning: Lessons from political philosophy. In Conference on fairness, accountability and transparency. PMLR, 149–159.
- Reuben Binns. 2020. On the apparent conflict between individual and group fairness. In Proceedings of the 2020 conference on fairness, accountability, and transparency. 514–524.
- Multimodal datasets: misogyny, pornography, and malignant stereotypes. arXiv preprint arXiv:2110.01963 (2021).
- Language (Technology) is Power: A Critical Survey of “Bias” in NLP. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics, Online, 5454–5476. https://doi.org/10.18653/v1/2020.acl-main.485
- Man is to computer programmer as woman is to homemaker? debiasing word embeddings. In Advances in Neural Information Processing Systems. 4349–4357.
- Bias and fairness in multimodal machine learning: A case study of automated video interviews. In Proceedings of the 2021 International Conference on Multimodal Interaction. 268–277.
- Joy Buolamwini and Timnit Gebru. 2018. Gender shades: Intersectional accuracy disparities in commercial gender classification. In Conference on fairness, accountability and transparency. PMLR, 77–91.
- Rubi: Reducing unimodal biases for visual question answering. Advances in neural information processing systems 32 (2019).
- IndoNLG: Benchmark and Resources for Evaluating Indonesian Natural Language Generation. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics, Online and Punta Cana, Dominican Republic, 8875–8898. https://doi.org/10.18653/v1/2021.emnlp-main.699
- Semantics derived automatically from language corpora contain human-like biases. Science 356, 6334 (2017), 183–186.
- L Elisa Celis and Vijay Keswani. 2020. Implicit diversity in image summarization. Proceedings of the ACM on Human-Computer Interaction 4, CSCW2 (2020), 1–28.
- Generative pretraining from pixels. In International conference on machine learning. PMLR, 1691–1703.
- Big Self-Supervised Models are Strong Semi-Supervised Learners. arXiv preprint arXiv:2006.10029 (2020).
- Uniter: Universal image-text representation learning. In ECCV.
- FairFil: Contrastive Neural Debiasing Method for Pretrained Text Encoders. In International Conference on Learning Representations. https://openreview.net/forum?id=N6JECD-PI5w
- Unifying Vision-and-Language Tasks via Text Generation. In Proceedings of the 38th International Conference on Machine Learning (Proceedings of Machine Learning Research, Vol. 139), Marina Meila and Tong Zhang (Eds.). PMLR, 1931–1942. https://proceedings.mlr.press/v139/cho21a.html
- Dall-eval: Probing the reasoning skills and social biases of text-to-image generative transformers. arXiv preprint arXiv:2202.04053 (2022).
- Fair generative modeling via weak supervision. In International Conference on Machine Learning. PMLR, 1887–1898.
- Debiasing vision-language models via biased prompts. arXiv preprint arXiv:2302.00070 (2023).
- Ching-Yao Chuang and Youssef Mroueh. 2021. Fair mixup: Fairness via interpolation. arXiv preprint arXiv:2103.06503 (2021).
- Alexis Conneau and Guillaume Lample. 2019. Cross-Lingual Language Model Pretraining. Curran Associates Inc., Red Hook, NY, USA.
- Flexibly fair representation learning by disentanglement. In International conference on machine learning. PMLR, 1436–1445.
- Quantifying Social Biases in NLP: A Generalization and Empirical Comparison of Extrinsic Fairness Metrics. Transactions of the Association for Computational Linguistics 9 (2021), 1249–1267.
- Enabling Multimodal Generation on CLIP via Vision-Language Knowledge Distillation. In Findings of the Association for Computational Linguistics: ACL 2022. Association for Computational Linguistics, Dublin, Ireland, 2383–2395. https://doi.org/10.18653/v1/2022.findings-acl.187
- Does object recognition work for everyone?. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition workshops. 52–59.
- Measuring Fairness with Biased Rulers: A Comparative Study on Bias Metrics for Pre-trained Language Models. In Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. 1693–1706.
- Image counterfactual sensitivity analysis for detecting unintended bias. arXiv preprint arXiv:1906.06439 (2019).
- Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018).
- Queens are Powerful too: Mitigating Gender Bias in Dialogue Generation. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP). 8173–8188.
- Measuring and Mitigating Unintended Bias in Text Classification. In AAAI.
- Unified Language Model Pre-training for Natural Language Understanding and Generation. In Advances in Neural Information Processing Systems, H. Wallach, H. Larochelle, A. Beygelzimer, F. d'Alché-Buc, E. Fox, and R. Garnett (Eds.), Vol. 32. Curran Associates, Inc. https://proceedings.neurips.cc/paper/2019/file/c20bb2d9a50d5ac1f713f8b34d9aac5a-Paper.pdf
- Imbalanced deep learning by minority class incremental rectification. IEEE transactions on pattern analysis and machine intelligence 41, 6 (2018), 1367–1381.
- An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. In International Conference on Learning Representations. https://openreview.net/forum?id=YicbFdNTTy
- Fairness through awareness. In Proceedings of the 3rd innovations in theoretical computer science conference. 214–226.
- Decoupled classifiers for group-fair and efficient machine learning. In Conference on fairness, accountability and transparency. PMLR, 119–133.
- Harrison Edwards and Amos Storkey. 2015. Censoring representations with an adversary. arXiv preprint arXiv:1511.05897 (2015).
- Yanai Elazar and Yoav Goldberg. 2018. Adversarial Removal of Demographic Attributes from Text Data. arXiv preprint arXiv:1808.06640 (2018).
- A confidence-based approach for balancing fairness and accuracy. In Proceedings of the 2016 SIAM international conference on data mining. SIAM, 144–152.
- Thomas B Fitzpatrick. 1975. Soleil et peau. J. Med. Esthet. 2 (1975), 33–34.
- Luciano Floridi and Massimo Chiriatti. 2020. GPT-3: Its nature, scope, limits, and consequences. Minds and Machines 30 (2020), 681–694.
- An ethical framework for a good AI society: Opportunities, risks, principles, and recommendations. Ethics, governance, and policies in artificial intelligence (2021), 19–39.
- A Friendly Face: Do Text-to-Image Systems Rely on Stereotypes when the Input is Under-Specified? arXiv preprint arXiv:2302.07159 (2023).
- Fair Diffusion: Instructing Text-to-Image Generation Models on Fairness. arXiv preprint arXiv:2302.10893 (2023).
- A survey on bias in deep NLP. Applied Sciences 11, 7 (2021), 3184.
- Fairness-aware ranking in search & recommendation systems with application to linkedin talent search. In Proceedings of the 25th acm sigkdd international conference on knowledge discovery & data mining. 2221–2231.
- Intrinsic Bias Metrics Do Not Correlate with Application Bias. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers). 1926–1940.
- Hila Gonen and Yoav Goldberg. 2019. Lipstick on a Pig: Debiasing Methods Cover up Systematic Gender Biases in Word Embeddings But do not Remove Them. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers). 609–614.
- Jointly de-biasing face recognition and demographic attribute estimation. In Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part XXIX 16. Springer, 330–347.
- Mitigating face recognition bias via group adaptive classifier. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 3414–3424.
- Anirudh Goyal and Yoshua Bengio. 2022. Inductive biases for deep learning of higher-level cognition. Proceedings of the Royal Society A 478, 2266 (2022), 20210068.
- Fairness indicators for systematic assessments of visual feature extractors. In 2022 ACM Conference on Fairness, Accountability, and Transparency. 70–88.
- Measuring individual differences in implicit cognition: the implicit association test. Journal of personality and social psychology 74, 6 (1998), 1464.
- Bias correction of learned generative models using likelihood-free importance weighting. Advances in neural information processing systems 32 (2019).
- Wei Guo and Aylin Caliskan. 2021. Detecting emergent intersectional biases: Contextualized word embeddings contain a distribution of human-like biases. In Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society. 122–133.
- Mitigating Gender Bias in Distilled Language Models via Counterfactual Role Reversal. In Findings of the Association for Computational Linguistics: ACL 2022. Association for Computational Linguistics, Dublin, Ireland, 658–678. https://doi.org/10.18653/v1/2022.findings-acl.55
- Towards a critical race methodology in algorithmic fairness. In Proceedings of the 2020 conference on fairness, accountability, and transparency. 501–512.
- Equality of opportunity in supervised learning. In Advances in neural information processing systems. 3315–3323.
- Casual Conversations v2: Designing a large consent-driven dataset to measure algorithmic bias and robustness. arXiv preprint arXiv:2211.05809 (2022).
- Towards measuring fairness in ai: the casual conversations dataset. IEEE Transactions on Biometrics, Behavior, and Identity Science 4, 3 (2021), 324–332.
- Masked Autoencoders Are Scalable Vision Learners. 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2021), 15979–15988.
- Mask R-CNN. In 2017 IEEE International Conference on Computer Vision (ICCV). 2980–2988. https://doi.org/10.1109/ICCV.2017.322
- Bias in machine learning-what is it good for?. In International Workshop on New Foundations for Human-Centered AI (NeHuAI) co-located with 24th European Conference on Artificial Intelligence (ECAI 2020), Virtual (Santiago de Compostela, Spain), September 4, 2020. RWTH Aachen University, 3–10.
- Women also snowboard: Overcoming bias in captioning models. In Proceedings of the European conference on computer vision (ECCV). 771–787.
- Scaling Up Vision-Language Pre-Training for Image Captioning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). 17980–17989.
- Deep imbalanced learning for face recognition and attribute prediction. IEEE transactions on pattern analysis and machine intelligence 42, 11 (2019), 2781–2794.
- Scaling Up Visual and Vision-Language Representation Learning With Noisy Text Supervision. In Proceedings of the 38th International Conference on Machine Learning (Proceedings of Machine Learning Research, Vol. 139), Marina Meila and Tong Zhang (Eds.). PMLR, 4904–4916. https://proceedings.mlr.press/v139/jia21b.html
- Heinrich Jiang and Ofir Nachum. 2020. Identifying and correcting label bias in machine learning. In International Conference on Artificial Intelligence and Statistics. PMLR, 702–712.
- The global landscape of AI ethics guidelines. Nature machine intelligence 1, 9 (2019), 389–399.
- Jungseock Joo and Kimmo Kärkkäinen. 2020. Gender slopes: Counterfactual fairness for computer vision models by attribute manipulation. In Proceedings of the 2nd international workshop on fairness, accountability, transparency and ethics in multimedia. 1–5.
- Masahiro Kaneko and Danushka Bollegala. 2019. Gender-preserving Debiasing for Pre-trained Word Embeddings. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics, Florence, Italy, 1641–1650. https://doi.org/10.18653/v1/P19-1160
- Kimmo Karkkainen and Jungseock Joo. 2021. Fairface: Face attribute dataset for balanced race, gender, and age for bias measurement and mitigation. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision. 1548–1558.
- Unequal representation and gender stereotypes in image search results for occupations. In Proceedings of the 33rd Annual ACM Conference on Human Factors in Computing Systems. ACM, 3819–3828.
- Preventing fairness gerrymandering: Auditing and learning for subgroup fairness. In International conference on machine learning. PMLR, 2564–2572.
- An empirical study of rich subgroup fairness for machine learning. In Proceedings of the conference on fairness, accountability, and transparency. 100–109.
- Striking the right balance with uncertainty. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 103–112.
- Age bias in emotion detection: An analysis of facial emotion recognition performance on young, middle-aged, and older adults. In Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society. 638–644.
- Multiaccuracy: Black-box post-processing for fairness in classification. In Proceedings of the 2019 AAAI/ACM Conference on AI, Ethics, and Society. 247–254.
- ViLT: Vision-and-Language Transformer Without Convolution or Region Supervision. In Proceedings of the 38th International Conference on Machine Learning (Proceedings of Machine Learning Research, Vol. 139), Marina Meila and Tong Zhang (Eds.). PMLR, 5583–5594. http://proceedings.mlr.press/v139/kim21k.html
- Svetlana Kiritchenko and Saif Mohammad. 2018. Examining Gender and Race Bias in Two Hundred Sentiment Analysis Systems. In Proceedings of the Seventh Joint Conference on Lexical and Computational Semantics. 43–53.
- Counterfactual fairness. Advances in neural information processing systems 30 (2017).
- Exploring Social Bias in Chatbots using Stereotype Knowledge. In Proceedings of the 2019 Workshop on Widening NLP. 177–180.
- BLIP: Bootstrapping Language-Image Pre-training for Unified Vision-Language Understanding and Generation. In Proceedings of the 39th International Conference on Machine Learning (Proceedings of Machine Learning Research, Vol. 162), Kamalika Chaudhuri, Stefanie Jegelka, Le Song, Csaba Szepesvari, Gang Niu, and Sivan Sabato (Eds.). PMLR, 12888–12900. https://proceedings.mlr.press/v162/li22n.html
- Align before fuse: Vision and language representation learning with momentum distillation. Advances in neural information processing systems 34 (2021), 9694–9705.
- VISUALBERT: ASimple AND PERFORMANT BASELINE FOR VISION AND LANGUAGE. arXiv preprint arXiv:1908.03557 (2019).
- Oscar: Object-Semantics Aligned Pre-training for Vision-Language Tasks. ECCV 2020 (2020).
- Towards Robust and Privacy-preserving Text Representations. arXiv preprint arXiv:1805.06093 (2018).
- Scaling Language-Image Pre-training via Masking. ArXiv abs/2212.00794 (2022).
- Towards Debiasing Sentence Representations. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics, Online, 5502–5515. https://doi.org/10.18653/v1/2020.acl-main.488
- M6: Multi-Modality-to-Multi-Modality Multitask Mega-Transformer for Unified Pretraining. In Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining (Virtual Event, Singapore) (KDD ’21). Association for Computing Machinery, New York, NY, USA, 3251–3261. https://doi.org/10.1145/3447548.3467206
- Microsoft COCO: Common Objects in Context. In Computer Vision – ECCV 2014. Springer International Publishing, 740–755. https://doi.org/10.1007/978-3-319-10602-1_48
- Does Gender Matter? Towards Fairness in Dialogue Systems. In Proceedings of the 28th International Conference on Computational Linguistics. 4403–4416.
- Multilingual Denoising Pre-training for Neural Machine Translation. Transactions of the Association for Computational Linguistics 8 (2020), 726–742. https://doi.org/10.1162/tacl_a_00343
- On the fairness of disentangled representations. Advances in neural information processing systems 32 (2019).
- Fairalm: Augmented lagrangian method for training fair models with little regret. In Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part XII. Springer, 365–381.
- Vilbert: Pretraining task-agnostic visiolinguistic representations for vision-and-language tasks. In Advances in Neural Information Processing Systems. 13–23.
- Gender bias in neural natural language processing. Logic, Language, and Security: Essays Dedicated to Andre Scedrov on the Occasion of His 65th Birthday (2020), 189–202.
- Learning adversarially fair and transferable representations. In International Conference on Machine Learning. PMLR, 3384–3393.
- Black is to Criminal as Caucasian is to Police: Detecting and Removing Multiclass Bias in Word Embeddings. In Proceedings of NAACL-HLT. 615–621.
- It’s All in the Name: Mitigating Gender Bias with Name-Based Counterfactual Data Substitution. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP). 5267–5275.
- On Measuring Social Biases in Sentence Encoders. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers). 622–628.
- Characterizing bias in classifiers using generative models. Advances in neural information processing systems 32 (2019).
- An Empirical Survey of the Effectiveness of Debiasing Techniques for Pre-trained Language Models. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). 1878–1898.
- A survey on bias and fairness in machine learning. ACM Computing Surveys (CSUR) 54, 6 (2021), 1–35.
- Ellis P Monk Jr. 2014. Skin tone stratification among Black Americans, 2001–2003. Social Forces 92, 4 (2014), 1313–1337.
- Invariant representations without adversarial training. Advances in Neural Information Processing Systems 31 (2018).
- Semantic Specialisation of Distributional Word Vector Spaces using Monolingual and Cross-Lingual Constraints. Transactions of the Association for Computational Linguistics 5 ([n. d.]).
- Understanding unequal gender classification accuracy from face images. arXiv preprint arXiv:1812.00099 (2018).
- StereoSet: Measuring stereotypical bias in pretrained language models. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers). 5356–5371.
- CrowS-Pairs: A Challenge Dataset for Measuring Social Biases in Masked Language Models. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP). 1953–1967.
- French CrowS-Pairs: Extending a challenge dataset for measuring social bias in masked language models to a language other than English. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). Association for Computational Linguistics, Dublin, Ireland, 8521–8531. https://doi.org/10.18653/v1/2022.acl-long.583
- Equity in skin typing: why it is time to replace the Fitzpatrick scale. British Journal of Dermatology 185, 1 (2021), 198–199.
- Reducing Gender Bias in Abusive Language Detection. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics, Brussels, Belgium, 2799–2804. https://doi.org/10.18653/v1/D18-1302
- On fairness and calibration. Advances in neural information processing systems 30 (2017).
- Counterfactual Inference for Text Classification Debiasing. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers). Association for Computational Linguistics, Online, 5434–5445. https://doi.org/10.18653/v1/2021.acl-long.422
- Learning Transferable Visual Models From Natural Language Supervision. In Proceedings of the 38th International Conference on Machine Learning (Proceedings of Machine Learning Research, Vol. 139), Marina Meila and Tong Zhang (Eds.). PMLR, 8748–8763. https://proceedings.mlr.press/v139/radford21a.html
- Language models are unsupervised multitask learners. OpenAI blog 1, 8 (2019), 9.
- Fair attribute classification through latent space de-biasing. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 9301–9310.
- Zero-shot text-to-image generation. In International Conference on Machine Learning. PMLR, 8821–8831.
- Lauren Rhue. 2018. Racial influence on automated perceptions of emotions. Available at SSRN 3281765 (2018).
- The Dollar Street Dataset: Images Representing the Geographic and Socioeconomic Diversity of the World. In Thirty-sixth Conference on Neural Information Processing Systems Datasets and Benchmarks Track. https://openreview.net/forum?id=qnfYsave0U4
- High-Resolution Image Synthesis With Latent Diffusion Models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). 10684–10695.
- Measuring social biases in grounded vision and language embeddings. arXiv preprint arXiv:2002.08911 (2020).
- Wendy D. Roth. 2016. The multiple dimensions of race. Ethnic and Racial Studies 39, 8 (2016), 1310–1338. https://doi.org/10.1080/01419870.2016.1140793
- Gender Bias in Coreference Resolution. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 2 (Short Papers). Association for Computational Linguistics, New Orleans, Louisiana, 8–14. https://doi.org/10.18653/v1/N18-2002
- Inclusivefacenet: Improving face attribute detection with race and gender diversity. arXiv preprint arXiv:1712.00193 (2017).
- Chiheon Kim Doyup Lee Saehoon Kim, Sanghun Cho and Woonhyuk Baek. 2021. minDALL-E on Conceptual Captions. https://github.com/kakaobrain/minDALL-E.
- Distributionally robust neural networks for group shifts: On the importance of regularization for worst-case generalization. arXiv preprint arXiv:1911.08731 (2019).
- Masked Language Model Scoring. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics, Online, 2699–2712. https://doi.org/10.18653/v1/2020.acl-main.240
- Fairness GAN: Generating datasets with fairness properties using a generative adversarial network. IBM Journal of Research and Development 63, 4/5 (2019), 3–1.
- Understanding contrastive learning requires incorporating inductive biases. In International Conference on Machine Learning. PMLR, 19250–19286.
- Gender bias in machine translation. Transactions of the Association for Computational Linguistics 9 (2021), 845–874.
- How computers see gender: An evaluation of gender classification in commercial facial analysis services. Proceedings of the ACM on Human-Computer Interaction 3, CSCW (2019), 1–33.
- Self-Diagnosis and Self-Debiasing: A Proposal for Reducing Corpus-Based Bias in NLP. Transactions of the Association for Computational Linguistics 9 (12 2021), 1408–1424. https://doi.org/10.1162/tacl_a_00434 arXiv:https://direct.mit.edu/tacl/article-pdf/doi/10.1162/tacl_a_00434/1979270/tacl_a_00434.pdf
- A step toward more inclusive people annotations for fairness. In Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society. 916–925.
- DeAR: Debiasing Vision-Language Models with Additive Residuals. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 6820–6829.
- No classification without representation: Assessing geodiversity issues in open data sets for the developing world. arXiv preprint arXiv:1711.08536 (2017).
- Sanjay Sharma. 2020. Algorithms of oppression: how search engines reinforce racism: by Safiya Umoja Noble, New York, USA, NYU Press, 2018, xv+ 229pp.,£ 22.99 (paperback), ISBN 9781479837243.
- How Much Can CLIP Benefit Vision-and-Language Tasks?. In International Conference on Learning Representations. https://openreview.net/forum?id=zf_Ll3HZWgy
- Revealing persona biases in dialogue systems. arXiv preprint arXiv:2104.08728 (2021).
- The Woman Worked as a Babysitter: On Biases in Language Generation. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP). 3407–3412.
- Towards Controllable Biases in Language Generation. In Findings of the Association for Computational Linguistics: EMNLP 2020. 3239–3254.
- Societal Biases in Language Generation: Progress and Challenges. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers). 4275–4293.
- Face recognition algorithm bias: Performance differences on images of children and adults. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition workshops. 0–0.
- Tejas Srinivasan and Yonatan Bisk. 2021. Worst of both worlds: Biases compound in pre-trained vision-and-language models. arXiv preprint arXiv:2104.08666 (2021).
- Dropout: a simple way to prevent neural networks from overfitting. The journal of machine learning research 15, 1 (2014), 1929–1958.
- Ryan Steed and Aylin Caliskan. 2021. Image representations learned with unsupervised pre-training contain human-like biases. In Proceedings of the 2021 ACM conference on fairness, accountability, and transparency. 701–713.
- VL-BERT: Pre-training of Generic Visual-Linguistic Representations. In International Conference on Learning Representations. https://openreview.net/forum?id=SygXPaEYvH
- Mitigating gender bias in natural language processing: Literature review. arXiv preprint arXiv:1906.08976 (2019).
- Hao Tan and Mohit Bansal. 2019. LXMERT: Learning Cross-Modality Encoder Representations from Transformers. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP). Association for Computational Linguistics, Hong Kong, China, 5100–5111. https://doi.org/10.18653/v1/D19-1514
- Yi Chern Tan and L Elisa Celis. 2019. Assessing social and intersectional biases in contextualized word representations. In Proceedings of the 33rd International Conference on Neural Information Processing Systems. 13230–13241.
- Christopher TH Teo and Ngai-Man Cheung. 2021. Measuring fairness in generative models. arXiv preprint arXiv:2107.07754 (2021).
- Emiel Van Miltenburg. 2016. Stereotyping and bias in the flickr30k dataset. arXiv preprint arXiv:1605.06083 (2016).
- Getting Gender Right in Neural Machine Translation. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing. 3003–3008.
- Russell T Vought. 2021. White House guidance for regulation of Artificial Intelligence Applications. https://www.ai.gov/white-house-guidance-for-regulation-of-artificial-intelligence-applications/
- Are Gender-Neutral Queries Really Gender-Neutral? Mitigating Gender Bias in Image Search. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing. 1995–2008.
- Mei Wang and Weihong Deng. 2020. Mitigating bias in face recognition using skewness-aware reinforcement learning. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 9322–9331.
- Racial faces in the wild: Reducing racial bias by information maximization adaptation network. In Proceedings of the ieee/cvf international conference on computer vision. 692–702.
- Balanced datasets are not enough: Estimating and mitigating gender bias in deep image representations. In Proceedings of the IEEE/CVF International Conference on Computer Vision. 5310–5319.
- Towards fairness in visual recognition: Effective strategies for bias mitigation. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 8919–8928.
- SimVLM: Simple Visual Language Model Pretraining with Weak Supervision. In International Conference on Learning Representations. https://openreview.net/forum?id=GUrhfTuf_3
- Racial limitations of Fitzpatrick skin type. Cutis 105, 2 (2020), 77–80.
- Mind the GAP: A balanced corpus of gendered ambiguous pronouns. Transactions of the Association for Computational Linguistics 6 (2018), 605–617.
- Measuring and Reducing Gendered Correlations in Pre-trained Models. (2020).
- Cost-sensitive learning vs. sampling: Which is best for handling unbalanced classes with unequal error costs? Dmin 7, 35-41 (2007), 24.
- Predictive inequity in object detection. arXiv preprint arXiv:1902.11097 (2019).
- On Modality Bias in the TVQA Dataset. In Proceedings of the British Machine Vision Conference (BMVC).
- Robert Wolfe and Aylin Caliskan. 2022a. American== white in multimodal language-and-image ai. In Proceedings of the 2022 AAAI/ACM Conference on AI, Ethics, and Society. 800–812.
- Robert Wolfe and Aylin Caliskan. 2022b. Markedness in visual semantic ai. In 2022 ACM Conference on Fairness, Accountability, and Transparency. 1269–1279.
- BigScience Workshop. 2023. BLOOM: A 176B-Parameter Open-Access Multilingual Language Model. arXiv:2211.05100 [cs.CL]
- Fairgan: Fairness-aware generative adversarial networks. In 2018 IEEE International Conference on Big Data (Big Data). IEEE, 570–575.
- Unified Detoxifying and Debiasing in Language Generation via Inference-time Adaptive Optimization. In The Eleventh International Conference on Learning Representations. https://openreview.net/forum?id=FvevdI0aA_h
- FILIP: Fine-grained Interactive Language-Image Pre-Training. In International Conference on Learning Representations. https://openreview.net/forum?id=cpDhcsEDC2
- From image descriptions to visual denotations: New similarity metrics for semantic inference over event descriptions. Transactions of the Association for Computational Linguistics 2 (2014), 67–78. https://doi.org/10.1162/tacl_a_00166
- Learning fair representations. In International conference on machine learning. PMLR, 325–333.
- LiT: Zero-Shot Transfer with Locked-image Text Tuning. CVPR (2022).
- Mitigating unwanted biases with adversarial learning. Proceedings of AAAI/ACM Conference on Ethics and Society(AIES) 2018 (2018).
- Michael Zhang and Christopher Re. [n. d.]. Contrastive Adapters for Foundation Model Group Robustness. In Advances in Neural Information Processing Systems.
- VinVL: Making Visual Representations Matter in Vision-Language Models. CVPR 2021 (2021).
- Counterfactually Measuring and Eliminating Social Bias in Vision-Language Pre-training Models. In Proceedings of the 30th ACM International Conference on Multimedia. 4996–5004.
- Age progression/regression by conditional adversarial autoencoder. In Proceedings of the IEEE conference on computer vision and pattern recognition. 5810–5818.
- Gender Bias in Contextualized Word Embeddings. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers). Association for Computational Linguistics, Minneapolis, Minnesota, 629–634. https://doi.org/10.18653/v1/N19-1064
- Men also like shopping: Reducing gender bias amplification using corpus-level constraints. arXiv preprint arXiv:1707.09457 (2017).
- Gender Bias in Coreference Resolution: Evaluation and Debiasing Methods. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 2 (Short Papers). 15–20.
- Learning Gender-Neutral Word Embeddings. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics, Brussels, Belgium, 4847–4853. https://doi.org/10.18653/v1/D18-1521
- Unified Vision-Language Pre-Training for Image Captioning and VQA. Proceedings of the AAAI Conference on Artificial Intelligence 34, 07 (Apr. 2020), 13041–13049. https://doi.org/10.1609/aaai.v34i07.7005
- Counterfactual Data Augmentation for Mitigating Gender Stereotypes in Languages with Rich Morphology. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. 1651–1661.
- Nayeon Lee (28 papers)
- Yejin Bang (25 papers)
- Holy Lovenia (30 papers)
- Samuel Cahyawijaya (75 papers)
- Wenliang Dai (24 papers)
- Pascale Fung (151 papers)