Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Multilingual Machine Translation Systems from Microsoft for WMT21 Shared Task (2111.02086v1)

Published 3 Nov 2021 in cs.CL

Abstract: This report describes Microsoft's machine translation systems for the WMT21 shared task on large-scale multilingual machine translation. We participated in all three evaluation tracks including Large Track and two Small Tracks where the former one is unconstrained and the latter two are fully constrained. Our model submissions to the shared task were initialized with DeltaLM\footnote{\url{https://aka.ms/deltalm}}, a generic pre-trained multilingual encoder-decoder model, and fine-tuned correspondingly with the vast collected parallel data and allowed data sources according to track settings, together with applying progressive learning and iterative back-translation approaches to further improve the performance. Our final submissions ranked first on three tracks in terms of the automatic evaluation metric.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (11)
  1. Jian Yang (505 papers)
  2. Shuming Ma (83 papers)
  3. Haoyang Huang (27 papers)
  4. Dongdong Zhang (79 papers)
  5. Li Dong (154 papers)
  6. Shaohan Huang (79 papers)
  7. Alexandre Muzio (8 papers)
  8. Saksham Singhal (14 papers)
  9. Hany Hassan Awadalla (24 papers)
  10. Xia Song (38 papers)
  11. Furu Wei (291 papers)
Citations (44)

Summary

We haven't generated a summary for this paper yet.