Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

First Align, then Predict: Understanding the Cross-Lingual Ability of Multilingual BERT (2101.11109v1)

Published 26 Jan 2021 in cs.CL

Abstract: Multilingual pretrained LLMs have demonstrated remarkable zero-shot cross-lingual transfer capabilities. Such transfer emerges by fine-tuning on a task of interest in one language and evaluating on a distinct language, not seen during the fine-tuning. Despite promising results, we still lack a proper understanding of the source of this transfer. Using a novel layer ablation technique and analyses of the model's internal representations, we show that multilingual BERT, a popular multilingual LLM, can be viewed as the stacking of two sub-networks: a multilingual encoder followed by a task-specific language-agnostic predictor. While the encoder is crucial for cross-lingual transfer and remains mostly unchanged during fine-tuning, the task predictor has little importance on the transfer and can be reinitialized during fine-tuning. We present extensive experiments with three distinct tasks, seventeen typologically diverse languages and multiple domains to support our hypothesis.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Benjamin Muller (20 papers)
  2. Yanai Elazar (44 papers)
  3. Benoît Sagot (60 papers)
  4. Djamé Seddah (28 papers)
Citations (65)
Github Logo Streamline Icon: https://streamlinehq.com