Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A Survey on Low-Resource Neural Machine Translation (2107.04239v1)

Published 9 Jul 2021 in cs.CL and cs.LG

Abstract: Neural approaches have achieved state-of-the-art accuracy on machine translation but suffer from the high cost of collecting large scale parallel data. Thus, a lot of research has been conducted for neural machine translation (NMT) with very limited parallel data, i.e., the low-resource setting. In this paper, we provide a survey for low-resource NMT and classify related works into three categories according to the auxiliary data they used: (1) exploiting monolingual data of source and/or target languages, (2) exploiting data from auxiliary languages, and (3) exploiting multi-modal data. We hope that our survey can help researchers to better understand this field and inspire them to design better algorithms, and help industry practitioners to choose appropriate algorithms for their applications.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Rui Wang (996 papers)
  2. Xu Tan (164 papers)
  3. Renqian Luo (19 papers)
  4. Tao Qin (201 papers)
  5. Tie-Yan Liu (242 papers)
Citations (51)