Social Bias in Large Language Models For Bangla: An Empirical Study on Gender and Religious Bias (2407.03536v3)
Abstract: The rapid growth of LLMs has put forward the study of biases as a crucial field. It is important to assess the influence of different types of biases embedded in LLMs to ensure fair use in sensitive fields. Although there have been extensive works on bias assessment in English, such efforts are rare and scarce for a major language like Bangla. In this work, we examine two types of social biases in LLM generated outputs for Bangla language. Our main contributions in this work are: (1) bias studies on two different social biases for Bangla, (2) a curated dataset for bias measurement benchmarking and (3) testing two different probing techniques for bias detection in the context of Bangla. This is the first work of such kind involving bias assessment of LLMs for Bangla to the best of our knowledge. All our code and resources are publicly available for the progress of bias related research in Bangla NLP.
- Persistent anti-muslim bias in large language models. Preprint, arXiv:2101.05783.
- AI@Meta. 2024. Llama 3 model card.
- BanglaParaphrase: A high-quality Bangla paraphrase dataset. In Proceedings of the 2nd Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics and the 12th International Joint Conference on Natural Language Processing (Volume 2: Short Papers), pages 261–272, Online only. Association for Computational Linguistics.
- CrossSum: Beyond English-centric cross-lingual summarization for 1,500+ language pairs. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 2541–2564, Toronto, Canada. Association for Computational Linguistics.
- Language (technology) is power: A critical survey of “bias” in NLP. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 5454–5476, Online. Association for Computational Linguistics.
- Man is to computer programmer as woman is to homemaker? debiasing word embeddings. Preprint, arXiv:1607.06520.
- Toward cultural bias evaluation datasets: The case of Bengali gender, religious, and national identity. In Proceedings of the First Workshop on Cross-Cultural Considerations in NLP (C3NLP), pages 68–83, Dubrovnik, Croatia. Association for Computational Linguistics.
- Toxicity in chatgpt: Analyzing persona-assigned language models. Preprint, arXiv:2304.05335.
- Disclosure and mitigation of gender bias in llms. Preprint, arXiv:2402.11190.
- Self-collaboration code generation via chatgpt. Preprint, arXiv:2304.07590.
- Certifying and removing disparate impact. Preprint, arXiv:1412.3756.
- Batya Friedman and Helen Nissenbaum. 1996. Bias in computer systems. ACM Trans. Inf. Syst., 14(3):330–347.
- Self-debiasing large language models: Zero-shot recognition and reduction of stereotypes. Preprint, arXiv:2402.01981.
- Mitigating gender bias in distilled language models via counterfactual role reversal. In Findings of the Association for Computational Linguistics: ACL 2022, pages 658–678, Dublin, Ireland. Association for Computational Linguistics.
- XL-sum: Large-scale multilingual abstractive summarization for 44 languages. In Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021, pages 4693–4703, Online. Association for Computational Linguistics.
- Uncovering implicit gender bias in narratives through commonsense inference. In Findings of the Association for Computational Linguistics: EMNLP 2021, pages 3866–3873, Punta Cana, Dominican Republic. Association for Computational Linguistics.
- A psychological study on the differences in attitude toward oppression among different generations of adult women in west bengal. International Journal of Indian Psychology, 9(4):144–150. DIP:18.01.014.20210904.
- Mistral 7b. Preprint, arXiv:2310.06825.
- The state and fate of linguistic diversity and inclusion in the NLP world. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 6282–6293, Online. Association for Computational Linguistics.
- Chatgpt for good? on opportunities and challenges of large language models for education. Learning and Individual Differences, 103:102274.
- Gender bias and stereotypes in large language models. In Proceedings of The ACM Collective Intelligence Conference, CI ’23. ACM.
- Gender and representation bias in GPT-3 generated stories. In Proceedings of the Third Workshop on Narrative Understanding, pages 48–55, Virtual. Association for Computational Linguistics.
- A survey on bias and fairness in machine learning. ACM Comput. Surv., 54(6).
- Aristides Milios and Parishad BehnamGhader. 2022. An analysis of social biases present in bert variants across multiple languages. Preprint, arXiv:2211.14402.
- Stereoset: Measuring stereotypical bias in pretrained language models. Preprint, arXiv:2004.09456.
- Crows-pairs: A challenge dataset for measuring social biases in masked language models. Preprint, arXiv:2010.00133.
- Biases in large language models: Origins, inventory, and discussion. J. Data and Information Quality, 15(2).
- Odiagenai: Generative ai and llm initiative for the odia language. https://github.com/shantipriyap/OdiaGenAI.
- A trip towards fairness: Bias and de-biasing in large language models. Preprint, arXiv:2305.13862.
- Gender bias in coreference resolution. CoRR, abs/1804.09301.
- An empirical study on the characteristics of bias upon context length variation for bangla. Preprint, arXiv:2406.17375.
- The woman worked as a babysitter: On biases in language generation. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pages 3407–3412, Hong Kong, China. Association for Computational Linguistics.
- Karolina Stanczak and Isabelle Augenstein. 2021. A survey on gender bias in natural language processing. Preprint, arXiv:2112.14168.
- Nishat Tarannum. 2019. A critical review on women oppression and threats in private spheres: Bangladesh perspective. American International Journal of Humanities, Arts and Social Sciences, 1:98–108.
- Llama 2: Open foundation and fine-tuned chat models. Preprint, arXiv:2307.09288.
- On evaluating and mitigating gender biases in multilingual settings. In Findings of the Association for Computational Linguistics: ACL 2023, pages 307–318, Toronto, Canada. Association for Computational Linguistics.
- Ethical and social risks of harm from language models. Preprint, arXiv:2112.04359.
- Personalizing dialogue agents: I have a dog, do you have pets too? In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 2204–2213, Melbourne, Australia. Association for Computational Linguistics.
- Gender bias in coreference resolution: Evaluation and debiasing methods. CoRR, abs/1804.06876.
- Gender bias in large language models across multiple languages. Preprint, arXiv:2403.00277.
- Jayanta Sadhu (4 papers)
- Maneesha Rani Saha (2 papers)
- Rifat Shahriyar (25 papers)