Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

TSMind: Alibaba and Soochow University's Submission to the WMT22 Translation Suggestion Task (2211.08987v1)

Published 16 Nov 2022 in cs.CL

Abstract: This paper describes the joint submission of Alibaba and Soochow University, TSMind, to the WMT 2022 Shared Task on Translation Suggestion (TS). We participate in the English-German and English-Chinese tasks. Basically, we utilize the model paradigm fine-tuning on the downstream tasks based on large-scale pre-trained models, which has recently achieved great success. We choose FAIR's WMT19 English-German news translation system and MBART50 for English-Chinese as our pre-trained models. Considering the task's condition of limited use of training data, we follow the data augmentation strategies proposed by WeTS to boost our TS model performance. The difference is that we further involve the dual conditional cross-entropy model and GPT-2 LLM to filter augmented data. The leader board finally shows that our submissions are ranked first in three of four language directions in the Naive TS task of the WMT22 Translation Suggestion task.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Xin Ge (4 papers)
  2. Ke Wang (531 papers)
  3. Jiayi Wang (74 papers)
  4. Nini Xiao (4 papers)
  5. Xiangyu Duan (10 papers)
  6. Yu Zhao (208 papers)
  7. Yuqi Zhang (54 papers)
Citations (2)