Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Understanding the Effect of Model Compression on Social Bias in Large Language Models (2312.05662v2)

Published 9 Dec 2023 in cs.CL
Understanding the Effect of Model Compression on Social Bias in Large Language Models

Abstract: LLMs trained with self-supervision on vast corpora of web text fit to the social biases of that text. Without intervention, these social biases persist in the model's predictions in downstream tasks, leading to representational harm. Many strategies have been proposed to mitigate the effects of inappropriate social biases learned during pretraining. Simultaneously, methods for model compression have become increasingly popular to reduce the computational burden of LLMs. Despite the popularity and need for both approaches, little work has been done to explore the interplay between these two. We perform a carefully controlled study of the impact of model compression via quantization and knowledge distillation on measures of social bias in LLMs. Longer pretraining and larger models led to higher social bias, and quantization showed a regularizer effect with its best trade-off around 20% of the original pretraining time.

Understanding the Effect of Model Compression on Social Bias in LLMs

The paper "Understanding the Effect of Model Compression on Social Bias in LLMs" investigates an underexplored area in the development of LLMs: the relationship between model compression techniques and the persistence of social biases within these models. This paper is pivotal for both theoretical understanding and practical advancements in the deployment of LLMs, as it bridges the gap between reducing computational burdens and mitigating biases.

Summary and Key Findings

LLMs are trained on massive datasets that inherently contain social biases. As these models evolve in size and capability, they tend to amplify these biases, thus creating challenges in ensuring fair and unbiased outputs in downstream tasks. This paper specifically focuses on two model compression techniques: quantization and knowledge distillation, examining their effects on social biases within BERT, RoBERTa, and Pythia models.

The research methodology utilizes a range of models of varying sizes and training extents to investigate the changes in bias metrics throughout the pretraining process. Evaluations are carried out using the Bias Bench benchmark, which encompasses three datasets focused on different social identities, allowing for a comprehensive examination of gender, race, and religion biases.

Notable findings from their experiments include:

  1. Compression as a Bias Mitigator: Both dynamic Post-Training Quantization (PTQ) and knowledge distillation exhibited potential in reducing social biases without significantly diminishing LLM (LM) performance. The paper underscores that while distillation often compromises LM efficacy, PTQ presents a more favorable balance between bias reduction and model capability preservation.
  2. Correlation Between Model Size and Bias: Larger model sizes and prolonged pretraining were associated with increased social biases. This correlation emphasizes the need for effective bias mitigation strategies, particularly as models scale.
  3. Inconsistent Bias Behavior Across Categories: The performance of models against different bias categories (gender, race, religion) varied. The inability of a singular model size to address all bias categories effectively reveals the nuanced nature of bias propagation in LLMs.
  4. Regularizing Effect of PTQ: Applying PTQ consistently across model types and sizes demonstrated a regularizing effect, suggesting its dual role in reducing both computational load and social biases.

Implications and Future Directions

This paper provides a dual lens on model compression as both an efficiency-enhancement tool and a prospective bias regulator. The immediate implications include advocating for PTQ as a standard practice in the deployment of LLMs, not just from a computational efficiency standpoint but also as a preemptive bias mitigation strategy.

The paper opens avenues for further research into more sophisticated compression techniques, such as pruning and adaptive computation, that might further decouple model efficiency from bias perpetuation. Additionally, refining bias evaluation datasets to include a broader array of biases, and extending them beyond Western-centric views, could provide a more holistic understanding of LLM fairness in global contexts.

Conclusion

This research offers a crucial contribution to the field of LLMs by addressing the intersection of model efficiency and ethical AI development. By showcasing the impact of compression techniques on social biases, it invites a rethinking of model deployment strategies in academia and industry. The findings serve as a foundation for bringing about balanced advancements in technology where model utility does not come at the cost of fair and unbiased language processing.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (34)
  1. The Low-Resource Double Bind: An Empirical Study of Pruning for Low-Resource Machine Translation. In EMNLP (Findings), pages 3316–3333. Association for Computational Linguistics.
  2. Intriguing Properties of Quantization at Scale. CoRR, abs/2305.19268.
  3. Pythia: A Suite for Analyzing Large Language Models Across Training and Scaling. CoRR, abs/2304.01373.
  4. Stereotyping Norwegian Salmon: An Inventory of Pitfalls in Fairness Benchmark Datasets. In ACL/IJCNLP (1), pages 1004–1015. Association for Computational Linguistics.
  5. The Lottery Ticket Hypothesis for Pre-trained BERT Networks. In NeurIPS.
  6. Pieter Delobelle and Bettina Berendt. 2022. FairDistillation: Mitigating Stereotyping in Language Models. In ECML/PKDD (2), volume 13714 of Lecture Notes in Computer Science, pages 638–654. Springer.
  7. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2019, Minneapolis, MN, USA, June 2-7, 2019, Volume 1 (Long and Short Papers), pages 4171–4186. Association for Computational Linguistics.
  8. Depth-Adaptive Transformer. In ICLR. OpenReview.net.
  9. The Pile: An 800GB Dataset of Diverse Text for Language Modeling. CoRR, abs/2101.00027.
  10. Debiasing Pre-Trained Language Models via Efficient Fine-Tuning. In LT-EDI, pages 59–69. Association for Computational Linguistics.
  11. Bridging Fairness and Environmental Sustainability in Natural Language Processing. In EMNLP, pages 7817–7836. Association for Computational Linguistics.
  12. Distilling the knowledge in a neural network. In NIPS Workshop on Deep Learning.
  13. What Do Compressed Deep Neural Networks Forget?
  14. Characterising Bias in Compressed Models. CoRR, abs/2010.03058.
  15. Masahiro Kaneko and Danushka Bollegala. 2021. Debiasing Pre-trained Contextualised Embeddings. In EACL, pages 1256–1266. Association for Computational Linguistics.
  16. Masahiro Kaneko and Danushka Bollegala. 2022. Unmasking the Mask - Evaluating Social Biases in Masked Language Models. In AAAI, pages 11954–11962. AAAI Press.
  17. Debiasing Isn’t Enough! - on the Effectiveness of Debiasing MLMs and Their Social Biases in Downstream Tasks. In COLING, pages 1299–1310. International Committee on Computational Linguistics.
  18. Beyond Distillation: Task-level Mixture-of-Experts for Efficient Inference. In EMNLP (Findings), pages 3577–3599. Association for Computational Linguistics.
  19. Measuring Bias in Contextualized Word Representations. CoRR, abs/1906.07337.
  20. When Do Pre-Training Biases Propagate to Downstream Tasks? A Case Study in Text Summarization. In EACL, pages 3198–3211. Association for Computational Linguistics.
  21. Towards Debiasing Sentence Representations. In ACL, pages 5502–5515. Association for Computational Linguistics.
  22. Lost in Pruning: The Effects of Pruning Neural Networks beyond Test Accuracy. In MLSys. mlsys.org.
  23. RoBERTa: A Robustly Optimized BERT Pretraining Approach. CoRR, abs/1907.11692.
  24. An Empirical Survey of the Effectiveness of Debiasing Techniques for Pre-trained Language Models. In ACL (1), pages 1878–1898. Association for Computational Linguistics.
  25. StereoSet: Measuring stereotypical bias in pretrained language models. In ACL/IJCNLP (1), pages 5356–5371. Association for Computational Linguistics.
  26. CrowS-Pairs: A Challenge Dataset for Measuring Social Biases in Masked Language Models. In EMNLP (1), pages 1953–1967. Association for Computational Linguistics.
  27. Intriguing Properties of Compression on Multilingual Models. In EMNLP, pages 9092–9110. Association for Computational Linguistics.
  28. Language models are unsupervised multitask learners.
  29. Null It Out: Guarding Protected Attributes by Iterative Nullspace Projection. In ACL, pages 7237–7256. Association for Computational Linguistics.
  30. DistilBERT, a distilled version of BERT: Smaller, faster, cheaper and lighter. CoRR, abs/1910.01108.
  31. Self-Diagnosis and Self-Debiasing: A Proposal for Reducing Corpus-Based Bias in NLP. Trans. Assoc. Comput. Linguistics, 9:1408–1424.
  32. Measuring and Reducing Gendered Correlations in Pre-trained Models. CoRR, abs/2010.06032.
  33. Beyond Preserved Accuracy: Evaluating Loyalty and Robustness of BERT Compression. In EMNLP (1), pages 10653–10659. Association for Computational Linguistics.
  34. Guangxuan Xu and Qingyuan Hu. 2022. Can Model Compression Improve NLP Fairness. CoRR, abs/2201.08542.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (2)
  1. Gustavo Gonçalves (1 paper)
  2. Emma Strubell (60 papers)
Citations (7)
Youtube Logo Streamline Icon: https://streamlinehq.com