2000 character limit reached
Exploring the traditional NMT model and Large Language Model for chat translation (2409.16331v1)
Published 24 Sep 2024 in cs.CL and cs.AI
Abstract: This paper describes the submissions of Huawei Translation Services Center(HW-TSC) to WMT24 chat translation shared task on English$\leftrightarrow$Germany (en-de) bidirection. The experiments involved fine-tuning models using chat data and exploring various strategies, including Minimum Bayesian Risk (MBR) decoding and self-training. The results show significant performance improvements in certain directions, with the MBR self-training method achieving the best results. The LLM also discusses the challenges and potential avenues for further research in the field of chat translation.
Collections
Sign up for free to add this paper to one or more collections.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.