Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

DiTTO: A Feature Representation Imitation Approach for Improving Cross-Lingual Transfer (2303.02357v1)

Published 4 Mar 2023 in cs.CL and cs.AI

Abstract: Zero-shot cross-lingual transfer is promising, however has been shown to be sub-optimal, with inferior transfer performance across low-resource languages. In this work, we envision languages as domains for improving zero-shot transfer by jointly reducing the feature incongruity between the source and the target language and increasing the generalization capabilities of pre-trained multilingual transformers. We show that our approach, DiTTO, significantly outperforms the standard zero-shot fine-tuning method on multiple datasets across all languages using solely unlabeled instances in the target language. Empirical results show that jointly reducing feature incongruity for multiple target languages is vital for successful cross-lingual transfer. Moreover, our model enables better cross-lingual transfer than standard fine-tuning methods, even in the few-shot setting.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Shanu Kumar (14 papers)
  2. Abbaraju Soujanya (1 paper)
  3. Sandipan Dandapat (17 papers)
  4. Sunayana Sitaram (54 papers)
  5. Monojit Choudhury (66 papers)