Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Understanding Cross-Lingual Syntactic Transfer in Multilingual Recurrent Neural Networks (2003.14056v3)

Published 31 Mar 2020 in cs.CL and cs.LG

Abstract: It is now established that modern neural LLMs can be successfully trained on multiple languages simultaneously without changes to the underlying architecture. But what kind of knowledge is really shared among languages within these models? Does multilingual training mostly lead to an alignment of the lexical representation spaces or does it also enable the sharing of purely grammatical knowledge? In this paper we dissect different forms of cross-lingual transfer and look for its most determining factors, using a variety of models and probing tasks. We find that exposing our LMs to a related language does not always increase grammatical knowledge in the target language, and that optimal conditions for lexical-semantic transfer may not be optimal for syntactic transfer.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (2)
  1. Prajit Dhar (4 papers)
  2. Arianna Bisazza (43 papers)
Citations (8)