Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

It's not Greek to mBERT: Inducing Word-Level Translations from Multilingual BERT (2010.08275v1)

Published 16 Oct 2020 in cs.CL

Abstract: Recent works have demonstrated that multilingual BERT (mBERT) learns rich cross-lingual representations, that allow for transfer across languages. We study the word-level translation information embedded in mBERT and present two simple methods that expose remarkable translation capabilities with no fine-tuning. The results suggest that most of this information is encoded in a non-linear way, while some of it can also be recovered with purely linear tools. As part of our analysis, we test the hypothesis that mBERT learns representations which contain both a language-encoding component and an abstract, cross-lingual component, and explicitly identify an empirical language-identity subspace within mBERT representations.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Hila Gonen (30 papers)
  2. Shauli Ravfogel (38 papers)
  3. Yanai Elazar (44 papers)
  4. Yoav Goldberg (142 papers)
Citations (47)