Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Training Neural Machine Translation To Apply Terminology Constraints (1906.01105v2)

Published 3 Jun 2019 in cs.CL, cs.AI, and cs.LG

Abstract: This paper proposes a novel method to inject custom terminology into neural machine translation at run time. Previous works have mainly proposed modifications to the decoding algorithm in order to constrain the output to include run-time-provided target terms. While being effective, these constrained decoding methods add, however, significant computational overhead to the inference step, and, as we show in this paper, can be brittle when tested in realistic conditions. In this paper we approach the problem by training a neural MT system to learn how to use custom terminology when provided with the input. Comparative experiments show that our method is not only more effective than a state-of-the-art implementation of constrained decoding, but is also as fast as constraint-free decoding.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Georgiana Dinu (17 papers)
  2. Prashant Mathur (21 papers)
  3. Marcello Federico (38 papers)
  4. Yaser Al-Onaizan (20 papers)
Citations (154)