Measuring Social Biases in Masked Language Models by Proxy of Prediction Quality (2402.13954v2)
Abstract: Innovative transformer-based LLMs produce contextually-aware token embeddings and have achieved state-of-the-art performance for a variety of natural language tasks, but have been shown to encode unwanted biases for downstream applications. In this paper, we evaluate the social biases encoded by transformers trained with the masked LLMing objective using proposed proxy functions within an iterative masking experiment to measure the quality of transformer models' predictions, and assess the preference of MLMs towards disadvantaged and advantaged groups. We compare bias estimations with those produced by other evaluation methods using benchmark datasets and assess their alignment with human annotated biases. We find relatively high religious and disability biases across considered MLMs and low gender bias in one dataset relative to another. We extend on previous work by evaluating social biases introduced after retraining an MLM under the masked LLMing objective, and find that proposed measures produce more accurate estimations of biases introduced by retraining MLMs than others based on relative preference for biased sentences between models.
- Man is to computer programmer as woman is to homemaker? debiasing word embeddings. In Advances in Neural Information Processing Systems, volume 29. Curran Associates, Inc.
- Language models are few-shot learners.
- BERT: pre-training of deep bidirectional transformers for language understanding. CoRR, abs/1810.04805.
- Bert: Pre-training of deep bidirectional transformers for language understanding.
- Understanding undesirable word embedding associations. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pages 1696–1705, Florence, Italy. Association for Computational Linguistics.
- Aaron Gokaslan and Vanya Cohen. 2019. Openwebtext corpus. http://Skylion007.github.io/OpenWebTextCorpus.
- Semantics derived automatically from language corpora necessarily contain human biases. CoRR, abs/1608.07187.
- Masahiro Kaneko and Danushka Bollegala. 2022. Unmasking the mask – evaluating social biases in masked language models. In Proceedings of the 36th AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada.
- Monolingual and multilingual reduction of gender bias in contextualized representations. In Proceedings of the 28th International Conference on Computational Linguistics, pages 5082–5093, Barcelona, Spain (Online). International Committee on Computational Linguistics.
- Improving multi-task deep neural networks via knowledge distillation for natural language understanding.
- Roberta: A robustly optimized BERT pretraining approach. CoRR, abs/1907.11692.
- Black is to criminal as Caucasian is to police: Detecting and removing multiclass bias in word embeddings. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 615–621, Minneapolis, Minnesota. Association for Computational Linguistics.
- On measuring social biases in sentence encoders.
- Quinn McNemar. 1947. Note on the sampling error of the difference between correlated proportions or percentages. Psychometrika, 12(2):153–157.
- Efficient estimation of word representations in vector space.
- Stereoset: Measuring stereotypical bias in pretrained language models.
- CrowS-pairs: A challenge dataset for measuring social biases in masked language models. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 1953–1967, Online. Association for Computational Linguistics.
- GloVe: Global vectors for word representation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 1532–1543, Doha, Qatar. Association for Computational Linguistics.
- Language models are unsupervised multitask learners.
- Masked language model scoring. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics.
- Quantifying the bias of transformer-based language models for african american english in masked language modeling. The Pacific-Asia Conference on Knowledge Discovery and Data Mining (PAKDD) 2023.
- Distilbert, a distilled version of BERT: smaller, faster, cheaper and lighter. CoRR, abs/1910.01108.
- S. S. Shapiro and M. B. Wilk. 1965. An analysis of variance test for normality (complete samples). Biometrika, 52(3/4):591–611.
- Bernard L Welch. 1947. The generalization of student’s’ problem when several different population variances are involved. Biometrika, pages 1–25.
- Huggingface’s transformers: State-of-the-art natural language processing. CoRR, abs/1910.03771.
- Aligning books and movies: Towards story-like visual explanations by watching movies and reading books. CoRR, abs/1506.06724.
- Rahul Zalkikar (1 paper)
- Kanchan Chandra (1 paper)