STACL: Simultaneous Translation with Implicit Anticipation and Controllable Latency using Prefix-to-Prefix Framework (1810.08398v5)
Abstract: Simultaneous translation, which translates sentences before they are finished, is useful in many scenarios but is notoriously difficult due to word-order differences. While the conventional seq-to-seq framework is only suitable for full-sentence translation, we propose a novel prefix-to-prefix framework for simultaneous translation that implicitly learns to anticipate in a single translation model. Within this framework, we present a very simple yet surprisingly effective wait-k policy trained to generate the target sentence concurrently with the source sentence, but always k words behind. Experiments show our strategy achieves low latency and reasonable quality (compared to full-sentence translation) on 4 directions: zh<->en and de<->en.
- Mingbo Ma (32 papers)
- Liang Huang (108 papers)
- Hao Xiong (41 papers)
- Renjie Zheng (29 papers)
- Kaibo Liu (17 papers)
- Baigong Zheng (19 papers)
- Chuanqiang Zhang (3 papers)
- Zhongjun He (19 papers)
- Hairong Liu (26 papers)
- Xing Li (82 papers)
- Hua Wu (191 papers)
- Haifeng Wang (194 papers)