The Unequal Opportunities of Large Language Models: Revealing Demographic Bias through Job Recommendations (2308.02053v2)
Abstract: LLMs have seen widespread deployment in various real-world applications. Understanding these biases is crucial to comprehend the potential downstream consequences when using LLMs to make decisions, particularly for historically disadvantaged groups. In this work, we propose a simple method for analyzing and comparing demographic bias in LLMs, through the lens of job recommendations. We demonstrate the effectiveness of our method by measuring intersectional biases within ChatGPT and LLaMA, two cutting-edge LLMs. Our experiments primarily focus on uncovering gender identity and nationality bias; however, our method can be extended to examine biases associated with any intersection of demographic identities. We identify distinct biases in both models toward various demographic identities, such as both models consistently suggesting low-paying jobs for Mexican workers or preferring to recommend secretarial roles to women. Our study highlights the importance of measuring the bias of LLMs in downstream applications to understand the potential for harm and inequitable outcomes.
- Persistent Anti-Muslim Bias in Large Language Models. In Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society (Virtual Event, USA) (AIES ’21). Association for Computing Machinery, New York, NY, USA, 298–306. https://doi.org/10.1145/3461702.3462624
- Heather Antecol and Kelly Bedard. 2004. The racial wage gap: The importance of labor force attachment differences across black, Mexican, and white men. Journal of Human Resources 39, 2 (2004), 564–583.
- Bard 2023. Google AI Updates: Bard and New AI Features in Search. Retrieved May 7, 2023 from https://blog.google/technology/ai/bard-google-ai-search-updates/
- Language (Technology) is Power: A Critical Survey of “Bias” in NLP. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics, Online, 5454–5476. https://doi.org/10.18653/v1/2020.acl-main.485
- Stereotyping Norwegian Salmon: An Inventory of Pitfalls in Fairness Benchmark Datasets. In Annual Meeting of the Association for Computational Linguistics.
- Semantics derived automatically from language corpora contain human-like biases. Science 356, 6334 (2017), 183–186. https://doi.org/10.1126/science.aal4230 arXiv:https://www.science.org/doi/pdf/10.1126/science.aal4230
- ChatGPT 2023. Introducing ChatGPT. Retrieved May 7, 2023 from https://openai.com/blog/chatgpt
- Crawling The Internal Knowledge-Base of Language Models. In Findings of the Association for Computational Linguistics: EACL 2023. Association for Computational Linguistics, Dubrovnik, Croatia, 1856–1869. https://aclanthology.org/2023.findings-eacl.139
- Emilio Ferrara. 2023. Should ChatGPT be Biased? Challenges and Risks of Bias in Large Language Models. arXiv:2304.03738 [cs.CY]
- Maarten Grootendorst. 2022. BERTopic: Neural topic modeling with a class-based TF-IDF procedure. arXiv:2203.05794 [cs.CL]
- Online negative sentiment towards Mexicans and Hispanics and impact on mental well-being: A time-series analysis of social media data during the 2016 United States presidential election. Heliyon 6, 9 (2020).
- HuggingChat 2023. HuggingChat. Retrieved May 7, 2023 from https://huggingface.co/chat/
- HuggingFace 2022. sentence-transformers/all-MiniLM-L6-v2. Retrieved May 7, 2023 from https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2
- Bias Out-of-the-Box: An Empirical Analysis of Intersectional Occupational Biases in Popular Generative Language Models. arXiv:2102.04130 [cs.CL]
- Li Lucy and David Bamman. 2021. Gender and Representation Bias in GPT-3 Generated Stories. In Proceedings of the Third Workshop on Narrative Understanding. Association for Computational Linguistics, Virtual, 48–55. https://doi.org/10.18653/v1/2021.nuse-1.5
- A Holistic Approach to Undesired Content Detection in the Real World. arXiv:2208.03274 [cs.CL]
- Douglas S Massey. 2009. Racial formation in theory and practice: The case of Mexicans in the United States. Race and social problems 1 (2009), 12–26.
- Robert W. McGee. 2023. Is Chat Gpt Biased Against Conservatives? An Empirical Study. (15 February 2023). https://doi.org/10.2139/ssrn.4359405
- UMAP: Uniform Manifold Approximation and Projection for Dimension Reduction. arXiv:1802.03426 [stat.ML]
- A Survey on Bias and Fairness in Machine Learning. ACM Comput. Surv. 54, 6, Article 115 (jul 2021), 35 pages. https://doi.org/10.1145/3457607
- StereoSet: Measuring stereotypical bias in pretrained language models. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers). Association for Computational Linguistics, Online, 5356–5371. https://doi.org/10.18653/v1/2021.acl-long.416
- Dissecting racial bias in an algorithm used to manage the health of populations. Science 366, 6464 (2019), 447–453. https://doi.org/10.1126/science.aax2342 arXiv:https://www.science.org/doi/pdf/10.1126/science.aax2342
- Orestis Papakyriakopoulos and Ethan Zuckerman. 2021. The media during the rise of trump: Identity politics, immigration,” Mexican” demonization and hate-crime. In Proceedings of the International AAAI Conference on Web and Social Media, Vol. 15. 467–478.
- On Natural Language User Profiles for Transparent and Scrutable Recommendation. In Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR ’22).
- Cordelia W. Reimers. 1983. Labor Market Discrimination Against Hispanic and Black Men. The Review of Economics and Statistics 65, 4 (1983), 570–579. http://www.jstor.org/stable/1935925
- David Rozado. 2023. The Political Biases of ChatGPT. Social Sciences 12, 3 (2023). https://doi.org/10.3390/socsci12030148
- The Self-Perception and Political Biases of ChatGPT. arXiv:2304.07333 [cs.CY]
- Quantifying Social Biases Using Templates is Unreliable. arXiv:2210.04337 [cs.CL]
- The Woman Worked as a Babysitter: On Biases in Language Generation. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP). Association for Computational Linguistics, Hong Kong, China, 3407–3412. https://doi.org/10.18653/v1/D19-1339
- LLaMA: Open and Efficient Foundation Language Models. arXiv:2302.13971 [cs.CL]
- Investigating Gender Bias in Language Models Using Causal Mediation Analysis. In Advances in Neural Information Processing Systems, H. Larochelle, M. Ranzato, R. Hadsell, M.F. Balcan, and H. Lin (Eds.), Vol. 33. Curran Associates, Inc., 12388–12401. https://proceedings.neurips.cc/paper_files/paper/2020/file/92650b2e92217715fe312e6fa7b90d82-Paper.pdf
- A prompt pattern catalog to enhance prompt engineering with chatgpt. arXiv preprint arXiv:2302.11382 (2023).
- Exploring AI Ethics of ChatGPT: A Diagnostic Analysis. arXiv:2301.12867 [cs.CL]
- Abel Salinas (5 papers)
- Parth Vipul Shah (2 papers)
- Yuzhong Huang (11 papers)
- Robert McCormack (3 papers)
- Fred Morstatter (64 papers)