Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Meta-Curriculum Learning for Domain Adaptation in Neural Machine Translation (2103.02262v1)

Published 3 Mar 2021 in cs.CL and cs.LG

Abstract: Meta-learning has been sufficiently validated to be beneficial for low-resource neural machine translation (NMT). However, we find that meta-trained NMT fails to improve the translation performance of the domain unseen at the meta-training stage. In this paper, we aim to alleviate this issue by proposing a novel meta-curriculum learning for domain adaptation in NMT. During meta-training, the NMT first learns the similar curricula from each domain to avoid falling into a bad local optimum early, and finally learns the curricula of individualities to improve the model robustness for learning domain-specific knowledge. Experimental results on 10 different low-resource domains show that meta-curriculum learning can improve the translation performance of both familiar and unfamiliar domains. All the codes and data are freely available at https://github.com/NLP2CT/Meta-Curriculum.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Runzhe Zhan (12 papers)
  2. Xuebo Liu (54 papers)
  3. Derek F. Wong (69 papers)
  4. Lidia S. Chao (41 papers)
Citations (43)
X Twitter Logo Streamline Icon: https://streamlinehq.com