Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Self-Augmented In-Context Learning for Unsupervised Word Translation (2402.10024v2)

Published 15 Feb 2024 in cs.CL, cs.AI, cs.IR, and cs.LG

Abstract: Recent work has shown that, while LLMs demonstrate strong word translation or bilingual lexicon induction (BLI) capabilities in few-shot setups, they still cannot match the performance of 'traditional' mapping-based approaches in the unsupervised scenario where no seed translation pairs are available, especially for lower-resource languages. To address this challenge with LLMs, we propose self-augmented in-context learning (SAIL) for unsupervised BLI: starting from a zero-shot prompt, SAIL iteratively induces a set of high-confidence word translation pairs for in-context learning (ICL) from an LLM, which it then reapplies to the same LLM in the ICL fashion. Our method shows substantial gains over zero-shot prompting of LLMs on two established BLI benchmarks spanning a wide range of language pairs, also outperforming mapping-based baselines across the board. In addition to achieving state-of-the-art unsupervised BLI performance, we also conduct comprehensive analyses on SAIL and discuss its limitations.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Yaoyiran Li (9 papers)
  2. Anna Korhonen (90 papers)
  3. Ivan Vulić (130 papers)
Citations (3)
X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets