Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Monotonic Simultaneous Translation with Chunk-wise Reordering and Refinement (2110.09646v1)

Published 18 Oct 2021 in cs.CL, cs.AI, and cs.LG

Abstract: Recent work in simultaneous machine translation is often trained with conventional full sentence translation corpora, leading to either excessive latency or necessity to anticipate as-yet-unarrived words, when dealing with a language pair whose word orders significantly differ. This is unlike human simultaneous interpreters who produce largely monotonic translations at the expense of the grammaticality of a sentence being translated. In this paper, we thus propose an algorithm to reorder and refine the target side of a full sentence translation corpus, so that the words/phrases between the source and target sentences are aligned largely monotonically, using word alignment and non-autoregressive neural machine translation. We then train a widely used wait-k simultaneous translation model on this reordered-and-refined corpus. The proposed approach improves BLEU scores and resulting translations exhibit enhanced monotonicity with source sentences.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. HyoJung Han (8 papers)
  2. Seokchan Ahn (1 paper)
  3. Yoonjung Choi (3 papers)
  4. Insoo Chung (3 papers)
  5. Sangha Kim (8 papers)
  6. Kyunghyun Cho (292 papers)
Citations (6)

Summary

We haven't generated a summary for this paper yet.