Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Larger-Scale Transformers for Multilingual Masked Language Modeling (2105.00572v1)

Published 2 May 2021 in cs.CL

Abstract: Recent work has demonstrated the effectiveness of cross-lingual LLM pretraining for cross-lingual understanding. In this study, we present the results of two larger multilingual masked LLMs, with 3.5B and 10.7B parameters. Our two new models dubbed XLM-R XL and XLM-R XXL outperform XLM-R by 1.8% and 2.4% average accuracy on XNLI. Our model also outperforms the RoBERTa-Large model on several English tasks of the GLUE benchmark by 0.3% on average while handling 99 more languages. This suggests pretrained models with larger capacity may obtain both strong performance on high-resource languages while greatly improving low-resource languages. We make our code and models publicly available.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Naman Goyal (37 papers)
  2. Jingfei Du (16 papers)
  3. Myle Ott (33 papers)
  4. Giri Anantharaman (2 papers)
  5. Alexis Conneau (33 papers)
Citations (112)