Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
131 tokens/sec
GPT-4o
10 tokens/sec
Gemini 2.5 Pro Pro
47 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

General Phrase Debiaser: Debiasing Masked Language Models at a Multi-Token Level (2311.13892v3)

Published 23 Nov 2023 in cs.CL and cs.AI

Abstract: The social biases and unwelcome stereotypes revealed by pretrained LLMs are becoming obstacles to their application. Compared to numerous debiasing methods targeting word level, there has been relatively less attention on biases present at phrase level, limiting the performance of debiasing in discipline domains. In this paper, we propose an automatic multi-token debiasing pipeline called \textbf{General Phrase Debiaser}, which is capable of mitigating phrase-level biases in masked LLMs. Specifically, our method consists of a \textit{phrase filter stage} that generates stereotypical phrases from Wikipedia pages as well as a \textit{model debias stage} that can debias models at the multi-token level to tackle bias challenges on phrases. The latter searches for prompts that trigger model's bias, and then uses them for debiasing. State-of-the-art results on standard datasets and metrics show that our approach can significantly reduce gender biases on both career and multiple disciplines, across models with varying parameter sizes.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (19)
  1. “Bert: Pre-training of deep bidirectional transformers for language understanding,” arXiv preprint arXiv:1810.04805, 2018.
  2. “Albert: A lite bert for self-supervised learning of language representations,” .
  3. “Roberta: A robustly optimized bert pretraining approach,” arXiv preprint arXiv:1907.11692, 2019.
  4. “Xlnet: Generalized autoregressive pretraining for language understanding,” in Advances in Neural Information Processing Systems, H. Wallach, H. Larochelle, A. Beygelzimer, F. d'Alché-Buc, E. Fox, and R. Garnett, Eds. 2019, vol. 32, Curran Associates, Inc.
  5. “Distilbert, a distilled version of bert: smaller, faster, cheaper and lighter,” arXiv preprint arXiv:1910.01108, 2019.
  6. “Learning transferable visual models from natural language supervision,” in International conference on machine learning. PMLR, 2021, pp. 8748–8763.
  7. “High-resolution image synthesis with latent diffusion models,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2022, pp. 10684–10695.
  8. “Towards debiasing sentence representations,” in Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, 2020, pp. 5502–5515.
  9. “Debiasing pre-trained contextualised embeddings,” in Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume, 2021, pp. 1256–1266.
  10. “He is very intelligent, she is very beautiful? on mitigating social biases in language modelling and generation,” in Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021, 2021, pp. 4534–4545.
  11. “An algorithm for the machine calculation of complex Fourier series,” Mathematics of Computation, vol. 19, no. 90, pp. 297–301, 1965.
  12. “Measuring and reducing gendered correlations in pre-trained models,” arXiv preprint arXiv:2010.06032, 2020.
  13. “Fairfil: Contrastive neural debiasing method for pretrained text encoders,” in International Conference on Learning Representations.
  14. “Auto-debias: Debiasing masked language models with automated biased prompts,” in Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2022, pp. 1012–1023.
  15. “How can we know what language models know?,” Transactions of the Association for Computational Linguistics, vol. 8, pp. 423–438, 2020.
  16. “On measuring social biases in sentence encoders,” in Proceedings of NAACL-HLT, 2019, pp. 622–628.
  17. “Beam search strategies for neural machine translation,” ACL 2017, p. 56, 2017.
  18. “Glue: A multi-task benchmark and analysis platform for natural language understanding,” in International Conference on Learning Representations.
  19. “Decoupled weight decay regularization,” arXiv preprint arXiv:1711.05101, 2017.
Citations (2)

Summary

We haven't generated a summary for this paper yet.