Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A Predictive Factor Analysis of Social Biases and Task-Performance in Pretrained Masked Language Models (2310.12936v2)

Published 19 Oct 2023 in cs.CL

Abstract: Various types of social biases have been reported with pretrained Masked LLMs (MLMs) in prior work. However, multiple underlying factors are associated with an MLM such as its model size, size of the training data, training objectives, the domain from which pretraining data is sampled, tokenization, and languages present in the pretrained corpora, to name a few. It remains unclear as to which of those factors influence social biases that are learned by MLMs. To study the relationship between model factors and the social biases learned by an MLM, as well as the downstream task performance of the model, we conduct a comprehensive study over 39 pretrained MLMs covering different model sizes, training objectives, tokenization methods, training data domains and languages. Our results shed light on important factors often neglected in prior literature, such as tokenization or model objectives.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Yi Zhou (438 papers)
  2. Jose Camacho-Collados (58 papers)
  3. Danushka Bollegala (84 papers)
Citations (5)

Summary

We haven't generated a summary for this paper yet.