Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Learning Kernel-Smoothed Machine Translation with Retrieved Examples (2109.09991v2)

Published 21 Sep 2021 in cs.CL

Abstract: How to effectively adapt neural machine translation (NMT) models according to emerging cases without retraining? Despite the great success of neural machine translation, updating the deployed models online remains a challenge. Existing non-parametric approaches that retrieve similar examples from a database to guide the translation process are promising but are prone to overfit the retrieved examples. In this work, we propose to learn Kernel-Smoothed Translation with Example Retrieval (KSTER), an effective approach to adapt neural machine translation models online. Experiments on domain adaptation and multi-domain machine translation datasets show that even without expensive retraining, KSTER is able to achieve improvement of 1.1 to 1.5 BLEU scores over the best existing online adaptation methods. The code and trained models are released at https://github.com/jiangqn/KSTER.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Qingnan Jiang (2 papers)
  2. Mingxuan Wang (83 papers)
  3. Jun Cao (108 papers)
  4. Shanbo Cheng (23 papers)
  5. Shujian Huang (106 papers)
  6. Lei Li (1293 papers)
Citations (32)