Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Efficient Wait-k Models for Simultaneous Machine Translation (2005.08595v2)

Published 18 May 2020 in cs.CL, cs.SD, and eess.AS

Abstract: Simultaneous machine translation consists in starting output generation before the entire input sequence is available. Wait-k decoders offer a simple but efficient approach for this problem. They first read k source tokens, after which they alternate between producing a target token and reading another source token. We investigate the behavior of wait-k decoding in low resource settings for spoken corpora using IWSLT datasets. We improve training of these models using unidirectional encoders, and training across multiple values of k. Experiments with Transformer and 2D-convolutional architectures show that our wait-k models generalize well across a wide range of latency levels. We also show that the 2D-convolution architecture is competitive with Transformers for simultaneous translation of spoken language.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Maha Elbayad (17 papers)
  2. Laurent Besacier (76 papers)
  3. Jakob Verbeek (59 papers)
Citations (73)

Summary

We haven't generated a summary for this paper yet.