Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
60 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

X-SNS: Cross-Lingual Transfer Prediction through Sub-Network Similarity (2310.17166v1)

Published 26 Oct 2023 in cs.CL

Abstract: Cross-lingual transfer (XLT) is an emergent ability of multilingual LLMs that preserves their performance on a task to a significant extent when evaluated in languages that were not included in the fine-tuning process. While English, due to its widespread usage, is typically regarded as the primary language for model adaption in various tasks, recent studies have revealed that the efficacy of XLT can be amplified by selecting the most appropriate source languages based on specific conditions. In this work, we propose the utilization of sub-network similarity between two languages as a proxy for predicting the compatibility of the languages in the context of XLT. Our approach is model-oriented, better reflecting the inner workings of foundation models. In addition, it requires only a moderate amount of raw text from candidate languages, distinguishing it from the majority of previous methods that rely on external resources. In experiments, we demonstrate that our method is more effective than baselines across diverse tasks. Specifically, it shows proficiency in ranking candidates for zero-shot XLT, achieving an improvement of 4.6% on average in terms of NDCG@3. We also provide extensive analyses that confirm the utility of sub-networks for XLT prediction.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Taejun Yun (2 papers)
  2. Jinhyeon Kim (5 papers)
  3. Deokyeong Kang (2 papers)
  4. Seong Hoon Lim (2 papers)
  5. Jihoon Kim (27 papers)
  6. Taeuk Kim (38 papers)