Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

IndiBias: A Benchmark Dataset to Measure Social Biases in Language Models for Indian Context (2403.20147v2)

Published 29 Mar 2024 in cs.CL

Abstract: The pervasive influence of social biases in language data has sparked the need for benchmark datasets that capture and evaluate these biases in LLMs. Existing efforts predominantly focus on English language and the Western context, leaving a void for a reliable dataset that encapsulates India's unique socio-cultural nuances. To bridge this gap, we introduce IndiBias, a comprehensive benchmarking dataset designed specifically for evaluating social biases in the Indian context. We filter and translate the existing CrowS-Pairs dataset to create a benchmark dataset suited to the Indian context in Hindi language. Additionally, we leverage LLMs including ChatGPT and InstructGPT to augment our dataset with diverse societal biases and stereotypes prevalent in India. The included bias dimensions encompass gender, religion, caste, age, region, physical appearance, and occupation. We also build a resource to address intersectional biases along three intersectional dimensions. Our dataset contains 800 sentence pairs and 300 tuples for bias measurement across different demographics. The dataset is available in English and Hindi, providing a size comparable to existing benchmark datasets. Furthermore, using IndiBias we compare ten different LLMs on multiple bias measurement metrics. We observed that the LLMs exhibit more bias across a majority of the intersectional groups.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (47)
  1. Navigating the social world: Toward an integrated framework for evaluating self, individuals, and groups. Psychological review.
  2. Persistent anti-muslim bias in large language models. In Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society, AIES ’21, page 298–306, New York, NY, USA. Association for Computing Machinery.
  3. B.R. Ambedkar. 2014. Annihilation of Caste: The Annotated Critical Edition. Verso Books.
  4. On the dangers of stochastic parrots: Can language models be too big? In Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency, FAccT ’21, page 610–623, New York, NY, USA. Association for Computing Machinery.
  5. Re-contextualizing fairness in NLP: The case of India. In Proceedings of the 2nd Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics and the 12th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 727–740, Online only. Association for Computational Linguistics.
  6. Language (technology) is power: A critical survey of “bias” in NLP. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 5454–5476, Online. Association for Computational Linguistics.
  7. Stereotyping Norwegian salmon: An inventory of pitfalls in fairness benchmark datasets. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 1004–1015, Online. Association for Computational Linguistics.
  8. Man is to computer programmer as woman is to homemaker? debiasing word embeddings.
  9. Semantics derived automatically from language corpora contain human-like biases. Science, 356(6334):183–186.
  10. Unsupervised cross-lingual representation learning at scale.
  11. Thomas A. de Souza. 1977. Regional and communal stereotypes of bombay university students. Indian Journal of Social Work, 38(1):37–44.
  12. Bernice: A multilingual pre-trained encoder for twitter. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, EMNLP 2022, Abu Dhabi, United Arab Emirates, December 7-11, 2022, pages 6191–6205. Association for Computational Linguistics.
  13. Towards leaving no indic language behind: Building monolingual corpora, benchmark and models for indic languages.
  14. M. Hammersley and R. Gomm. 1997. Bias in social research. Sociological Research Online, 2(1):7–19.
  15. Thongkholal Haokip. 2021. From ’chinky’ to ’coronavirus’: racism against northeast indians during the covid-19 pandemic. Asian Ethnicity, 22(2):353–373.
  16. Seegull: A stereotype benchmark with broad geo-cultural coverage leveraging generative models.
  17. Mitigating gender bias amplification in distribution by posterior regularization. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 2936–2942, Online. Association for Computational Linguistics.
  18. Mistral 7b.
  19. Muril: Multilingual representations for indian languages.
  20. Mascha Kurpicz-Briki. 2020. Cultural differences in bias? origin and gender bias in pre-trained german and french word embeddings.
  21. Benchmarking intersectional biases in NLP. In Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 3598–3609, Seattle, United States. Association for Computational Linguistics.
  22. AraWEAT: Multidimensional analysis of biases in Arabic word embeddings. In Proceedings of the Fifth Arabic Natural Language Processing Workshop.
  23. Socially aware bias measurements for hindi language representations.
  24. Black is to criminal as caucasian is to police: Detecting and removing multiclass bias in word embeddings. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies,.
  25. On measuring social biases in sentence encoders. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies.
  26. Duncan McDuie-Ra. 2012. Northeast migrants in Delhi: Race, refuge and retail. Amsterdam University Press.
  27. Hate speech detection and racial bias mitigation in social media based on bert model. PLOS ONE, 15:1–26.
  28. StereoSet: Measuring stereotypical bias in pretrained language models. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 5356–5371, Online. Association for Computational Linguistics.
  29. CrowS-pairs: A challenge dataset for measuring social biases in masked language models. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 1953–1967, Online. Association for Computational Linguistics.
  30. French CrowS-pairs: Extending a challenge dataset for measuring social bias in masked language models to a language other than English. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 8521–8531, Dublin, Ireland. Association for Computational Linguistics.
  31. Debiasing gender biased hindi words with word-embedding. In Proceedings of the 2019 2nd International Conference on Algorithms, Computing and Artificial Intelligence, ACAI ’19, page 450–456, New York, NY, USA. Association for Computing Machinery.
  32. Smart, responsible, and upper caste only: measuring caste attitudes through large-scale analysis of matrimonial profiles. In Proceedings of the International AAAI Conference on Web and Social Media, volume 13, pages 393–404.
  33. Nidhi Sabharwal and Wandana Sonalkar. 2015. Dalit women in india: At the crossroads of gender, class, and caste. Global justice: Theory, Practice, Rhetoric, 8.
  34. Re-imagining algorithmic fairness in india and beyond.
  35. Haspeede 2 @ evalita2020: Overview of the evalita 2020 hate speech detection task.
  36. Gender Bias in Machine Translation. Transactions of the Association for Computational Linguistics, 9:845–874.
  37. Societal biases in language generation: Progress and challenges.
  38. mgpt: Few-shot learners go multilingual.
  39. Can existing methods debias languages other than English? first attempt to analyze and mitigate Japanese word embeddings. In Proceedings of the Second Workshop on Gender Bias in Natural Language Processing, pages 44–55, Barcelona, Spain (Online). Association for Computational Linguistics.
  40. Yi Chern Tan and L. Elisa Celis. 2019. Assessing social and intersectional biases in contextualized word representations.
  41. No language left behind: Scaling human-centered machine translation.
  42. Llama 2: Open foundation and fine-tuned chat models.
  43. Huggingface’s transformers: State-of-the-art natural language processing.
  44. Bloom: A 176b-parameter open-access multilingual language model.
  45. mt5: A massively multilingual pre-trained text-to-text transformer.
  46. Men also like shopping: Reducing gender bias amplification using corpus-level constraints. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing.
  47. The moral integrity corpus: A benchmark for ethical dialogue systems. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 3755–3773, Dublin, Ireland. Association for Computational Linguistics.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Nihar Ranjan Sahoo (27 papers)
  2. Pranamya Prashant Kulkarni (1 paper)
  3. Narjis Asad (1 paper)
  4. Arif Ahmad (4 papers)
  5. Tanu Goyal (1 paper)
  6. Aparna Garimella (19 papers)
  7. Pushpak Bhattacharyya (153 papers)
Citations (5)

Summary

We haven't generated a summary for this paper yet.

Youtube Logo Streamline Icon: https://streamlinehq.com