Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

WeChat Neural Machine Translation Systems for WMT21 (2108.02401v2)

Published 5 Aug 2021 in cs.CL and cs.AI

Abstract: This paper introduces WeChat AI's participation in WMT 2021 shared news translation task on English->Chinese, English->Japanese, Japanese->English and English->German. Our systems are based on the Transformer (Vaswani et al., 2017) with several novel and effective variants. In our experiments, we employ data filtering, large-scale synthetic data generation (i.e., back-translation, knowledge distillation, forward-translation, iterative in-domain knowledge transfer), advanced finetuning approaches, and boosted Self-BLEU based model ensemble. Our constrained systems achieve 36.9, 46.9, 27.8 and 31.3 case-sensitive BLEU scores on English->Chinese, English->Japanese, Japanese->English and English->German, respectively. The BLEU scores of English->Chinese, English->Japanese and Japanese->English are the highest among all submissions, and that of English->German is the highest among all constrained submissions.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Xianfeng Zeng (5 papers)
  2. Yijin Liu (29 papers)
  3. Ernan Li (4 papers)
  4. Qiu Ran (5 papers)
  5. Fandong Meng (174 papers)
  6. Peng Li (390 papers)
  7. Jinan Xu (64 papers)
  8. Jie Zhou (687 papers)
Citations (20)

Summary

We haven't generated a summary for this paper yet.