Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
60 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A Simple and Effective Method to Improve Zero-Shot Cross-Lingual Transfer Learning (2210.09934v1)

Published 18 Oct 2022 in cs.CL

Abstract: Existing zero-shot cross-lingual transfer methods rely on parallel corpora or bilingual dictionaries, which are expensive and impractical for low-resource languages. To disengage from these dependencies, researchers have explored training multilingual models on English-only resources and transferring them to low-resource languages. However, its effect is limited by the gap between embedding clusters of different languages. To address this issue, we propose Embedding-Push, Attention-Pull, and Robust targets to transfer English embeddings to virtual multilingual embeddings without semantic loss, thereby improving cross-lingual transferability. Experimental results on mBERT and XLM-R demonstrate that our method significantly outperforms previous works on the zero-shot cross-lingual text classification task and can obtain a better multilingual alignment.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (9)
  1. Kunbo Ding (2 papers)
  2. Weijie Liu (33 papers)
  3. Yuejian Fang (18 papers)
  4. Weiquan Mao (7 papers)
  5. Zhe Zhao (97 papers)
  6. Tao Zhu (205 papers)
  7. Haoyan Liu (11 papers)
  8. Rong Tian (5 papers)
  9. Yiren Chen (13 papers)
Citations (7)