Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Data-Driven Adaptive Simultaneous Machine Translation (2204.12672v1)

Published 27 Apr 2022 in cs.CL

Abstract: In simultaneous translation (SimulMT), the most widely used strategy is the wait-k policy thanks to its simplicity and effectiveness in balancing translation quality and latency. However, wait-k suffers from two major limitations: (a) it is a fixed policy that can not adaptively adjust latency given context, and (b) its training is much slower than full-sentence translation. To alleviate these issues, we propose a novel and efficient training scheme for adaptive SimulMT by augmenting the training corpus with adaptive prefix-to-prefix pairs, while the training complexity remains the same as that of training full-sentence translation models. Experiments on two language pairs show that our method outperforms all strong baselines in terms of translation quality and latency.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (10)
  1. Guangxu Xun (2 papers)
  2. Mingbo Ma (32 papers)
  3. Yuchen Bian (5 papers)
  4. Xingyu Cai (10 papers)
  5. Jiaji Huang (17 papers)
  6. Renjie Zheng (29 papers)
  7. Junkun Chen (27 papers)
  8. Jiahong Yuan (12 papers)
  9. Kenneth Church (21 papers)
  10. Liang Huang (108 papers)