Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Discovering Language-neutral Sub-networks in Multilingual Language Models (2205.12672v2)

Published 25 May 2022 in cs.CL

Abstract: Multilingual pre-trained LLMs transfer remarkably well on cross-lingual downstream tasks. However, the extent to which they learn language-neutral representations (i.e., shared representations that encode similar phenomena across languages), and the effect of such representations on cross-lingual transfer performance, remain open questions. In this work, we conceptualize language neutrality of multilingual models as a function of the overlap between language-encoding sub-networks of these models. We employ the lottery ticket hypothesis to discover sub-networks that are individually optimized for various languages and tasks. Our evaluation across three distinct tasks and eleven typologically-diverse languages demonstrates that sub-networks for different languages are topologically similar (i.e., language-neutral), making them effective initializations for cross-lingual transfer with limited performance degradation.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Negar Foroutan (10 papers)
  2. Mohammadreza Banaei (8 papers)
  3. Remi Lebret (23 papers)
  4. Antoine Bosselut (85 papers)
  5. Karl Aberer (44 papers)
Citations (26)
X Twitter Logo Streamline Icon: https://streamlinehq.com