LLMs Are Zero-Shot Context-Aware Simultaneous Translators (2406.13476v3)
Abstract: The advent of transformers has fueled progress in machine translation. More recently LLMs have come to the spotlight thanks to their generality and strong performance in a wide range of language tasks, including translation. Here we show that open-source LLMs perform on par with or better than some state-of-the-art baselines in simultaneous machine translation (SiMT) tasks, zero-shot. We also demonstrate that injection of minimal background information, which is easy with an LLM, brings further performance gains, especially on challenging technical subject-matter. This highlights LLMs' potential for building next generation of massively multilingual, context-aware and terminologically accurate SiMT systems that require no resource-intensive training or fine-tuning.
- Roman Koshkin (3 papers)
- Katsuhito Sudoh (35 papers)
- Satoshi Nakamura (94 papers)