Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A Closer Look at Few-Shot Crosslingual Transfer: The Choice of Shots Matters (2012.15682v2)

Published 31 Dec 2020 in cs.CL

Abstract: Few-shot crosslingual transfer has been shown to outperform its zero-shot counterpart with pretrained encoders like multilingual BERT. Despite its growing popularity, little to no attention has been paid to standardizing and analyzing the design of few-shot experiments. In this work, we highlight a fundamental risk posed by this shortcoming, illustrating that the model exhibits a high degree of sensitivity to the selection of few shots. We conduct a large-scale experimental study on 40 sets of sampled few shots for six diverse NLP tasks across up to 40 languages. We provide an analysis of success and failure cases of few-shot transfer, which highlights the role of lexical features. Additionally, we show that a straightforward full model finetuning approach is quite effective for few-shot transfer, outperforming several state-of-the-art few-shot approaches. As a step towards standardizing few-shot crosslingual experimental designs, we make our sampled few shots publicly available.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Mengjie Zhao (35 papers)
  2. Yi Zhu (233 papers)
  3. Ehsan Shareghi (54 papers)
  4. Ivan Vulić (130 papers)
  5. Roi Reichart (82 papers)
  6. Anna Korhonen (90 papers)
  7. Hinrich Schütze (250 papers)
Citations (62)

Summary

We haven't generated a summary for this paper yet.