Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
60 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Dictionary-based Phrase-level Prompting of Large Language Models for Machine Translation (2302.07856v1)

Published 15 Feb 2023 in cs.CL and cs.LG
Dictionary-based Phrase-level Prompting of Large Language Models for Machine Translation

Abstract: LLMs demonstrate remarkable machine translation (MT) abilities via prompting, even though they were not explicitly trained for this task. However, even given the incredible quantities of data they are trained on, LLMs can struggle to translate inputs with rare words, which are common in low resource or domain transfer scenarios. We show that LLM prompting can provide an effective solution for rare words as well, by using prior knowledge from bilingual dictionaries to provide control hints in the prompts. We propose a novel method, DiPMT, that provides a set of possible translations for a subset of the input words, thereby enabling fine-grained phrase-level prompted control of the LLM. Extensive experiments show that DiPMT outperforms the baseline both in low-resource MT, as well as for out-of-domain MT. We further provide a qualitative analysis of the benefits and limitations of this approach, including the overall level of controllability that is achieved.

Dictionary-based Phrase-level Prompting of LLMs for Machine Translation

The paper "Dictionary-based Phrase-level Prompting of LLMs for Machine Translation" offers a novel approach to enhance the performance of LLMs in machine translation (MT), particularly addressing issues with rare word translations prevalent in low-resource or domain-specific scenarios. The authors present a method called DiPMT, which leverages bilingual dictionaries to provide phrase-level control hints within the prompts used for LLMs.

Summary of Approach

The researchers build on the already established capacity of LLMs to perform MT tasks via prompting without explicit training for translation. However, these models face significant challenges in translating rare words, a common issue in low-resource or domain transfer contexts. DiPMT addresses this shortcoming by incorporating dictionary knowledge directly into the prompts. This approach is inspired by previous success in supervised MT models that used dictionary information, but here, it is applied in a zero-shot fashion. The method enables LLMs to consider multiple potential translations for specified input words, thereby enhancing controllability at the word level.

This technique involves appending dictionary-based hints to the end of MT prompts, suggesting possible meanings for challenging words within the input. This method is designed to enrich the LLMs' knowledge of low-resource contexts without imposing hard constraints on the translations produced.

Experimental Results

The experimentation conducted by the authors spans multiple configurations involving LLMs like OPT and Bloom and languages selected from the Flores-101 dataset. The results underscore the ability of DiPMT to outperform baseline models in both low-resource and out-of-domain MT settings. Significant numerical improvements are reported, with BLEU score enhancements up to 13 points in particular domains. Moreover, analyses reveal that DiPMT maintains a level of controllability over translation outputs, which is a critical factor in improving MT quality.

Implications and Future Directions

From a practical perspective, DiPMT has the potential to improve translation quality in scenarios where access to comprehensive training data is limited. By integrating easily obtainable bilingual dictionaries into the prompting framework, DiPMT makes significant strides in translating rare and complex terminology. This carries substantial implications for contexts requiring specialized vocabulary, such as medical or technical domains.

Theoretically, this work highlights the potential of prompt engineering not only in tailoring LLM outputs but also in enriching them with structured external knowledge. The strategy of using dictionaries could be extrapolated to other tasks where domain-specific knowledge is crucial.

Looking ahead, further investigation could explore extending DiPMT to phrase-level translations beyond word-level hints. Moreover, enhancing the method of selecting dictionary entries and refining prompt structures could yield even greater improvements in translation accuracy and control, especially as more sophisticated bilingual and multilingual models are developed.

In conclusion, the proposed DiPMT method represents a noteworthy advancement in using prompting techniques to adapt LLMs effectively for machine translation, especially in scenarios involving low-resource language pairs and specialized domains.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Marjan Ghazvininejad (33 papers)
  2. Hila Gonen (30 papers)
  3. Luke Zettlemoyer (225 papers)
Citations (61)
Youtube Logo Streamline Icon: https://streamlinehq.com