Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Code-Switching for Enhancing NMT with Pre-Specified Translation (1904.09107v4)

Published 19 Apr 2019 in cs.CL

Abstract: Leveraging user-provided translation to constrain NMT has practical significance. Existing methods can be classified into two main categories, namely the use of placeholder tags for lexicon words and the use of hard constraints during decoding. Both methods can hurt translation fidelity for various reasons. We investigate a data augmentation method, making code-switched training data by replacing source phrases with their target translations. Our method does not change the MNT model or decoding algorithm, allowing the model to learn lexicon translations by copying source-side target words. Extensive experiments show that our method achieves consistent improvements over existing approaches, improving translation of constrained words without hurting unconstrained words.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Kai Song (21 papers)
  2. Yue Zhang (618 papers)
  3. Heng Yu (61 papers)
  4. Weihua Luo (63 papers)
  5. Kun Wang (355 papers)
  6. Min Zhang (630 papers)
Citations (112)