Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
125 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Language Model is a Branch Predictor for Simultaneous Machine Translation (2312.14488v1)

Published 22 Dec 2023 in cs.CL and cs.AI

Abstract: The primary objective of simultaneous machine translation (SiMT) is to minimize latency while preserving the quality of the final translation. Drawing inspiration from CPU branch prediction techniques, we propose incorporating branch prediction techniques in SiMT tasks to reduce translation latency. Specifically, we utilize a LLM as a branch predictor to predict potential branch directions, namely, future source words. Subsequently, we utilize the predicted source words to decode the output in advance. When the actual source word deviates from the predicted source word, we use the real source word to decode the output again, replacing the predicted output. To further reduce computational costs, we share the parameters of the encoder and the branch predictor, and utilize a pre-trained LLM for initialization. Our proposed method can be seamlessly integrated with any SiMT model. Extensive experimental results demonstrate that our approach can improve translation quality and latency at the same time. Our code is available at https://github.com/YinAoXiong/simt_branch_predictor .

Definition Search Book Streamline Icon: https://streamlinehq.com
References (23)
  1. “Can neural machine translation do simultaneous translation?,” 2016.
  2. “Learning to Translate in Real-time with Neural Machine Translation,” in EACL, 2017.
  3. “STACL: Simultaneous Translation with Implicit Anticipation and Controllable Latency using Prefix-to-Prefix Framework,” in ACL, 2019.
  4. “Monotonic Infinite Lookback Attention for Simultaneous Machine Translation,” in ACL, 2019.
  5. “Monotonic multihead attention,” in ICLR, 2020.
  6. James E. Smith, “A study of branch prediction strategies,” in 25 Years of the International Symposia on Computer Architecture (Selected Papers), 1998.
  7. “Efficient Wait-k Models for Simultaneous Machine Translation,” in Interspeech, 2020.
  8. “Modeling Dual Read/Write Paths for Simultaneous Machine Translation,” in ACL, 2022.
  9. “Information-Transport-based Policy for Simultaneous Translation,” 2022.
  10. “Learning Adaptive Segmentation Policy for Simultaneous Translation,” in EMNLP, 2020.
  11. “Mixspeech: Cross-modality self-learning with audio-visual stream mixup for visual speech translation and recognition,” in ICCV, 2023, pp. 15735–15745.
  12. “Opensr: Open-modality speech recognition via maintaining multi-modality alignment,” arXiv preprint arXiv:2306.06410, 2023.
  13. “Simulslt: End-to-end simultaneous sign language translation,” in ACM MM, 2021, pp. 4118–4127.
  14. “Mlslt: Towards multilingual sign language translation,” in CVPR, 2022, pp. 5109–5119.
  15. “Gloss attention for gloss-free sign language translation,” in CVPR, 2023, pp. 2551–2562.
  16. “Simullr: Simultaneous lip reading transducer with attention-guided adaptive memory,” in ACM MM, 2021, pp. 1359–1367.
  17. “The iwslt 2015 evaluation campaign,” 2015.
  18. “Findings of the 2015 workshop on statistical machine translation,” in Proceedings of the Tenth Workshop on Statistical Machine Translation, 2015.
  19. “Re-translation strategies for long form, simultaneous, spoken language translation,” in ICASSP, 2020.
  20. “Bleu: A Method for Automatic Evaluation of Machine Translation,” in Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics, 2002.
  21. “Language models are unsupervised multitask learners,” 2019.
  22. Sonia Vandepitte, “Anticipation in conference interpreting: A cognitive process,” Revista alicantina de estudios ingleses, No. 14 (Nov. 2001); pp. 323-335, 2001.
  23. “Towards Making the Most of Cross-Lingual Transfer for Zero-Shot Neural Machine Translation,” in ACL, 2022.
Citations (1)

Summary

We haven't generated a summary for this paper yet.