Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Unsupervised Neural Machine Translation for Low-Resource Domains via Meta-Learning (2010.09046v2)

Published 18 Oct 2020 in cs.CL, cs.AI, and cs.LG

Abstract: Unsupervised machine translation, which utilizes unpaired monolingual corpora as training data, has achieved comparable performance against supervised machine translation. However, it still suffers from data-scarce domains. To address this issue, this paper presents a novel meta-learning algorithm for unsupervised neural machine translation (UNMT) that trains the model to adapt to another domain by utilizing only a small amount of training data. We assume that domain-general knowledge is a significant factor in handling data-scarce domains. Hence, we extend the meta-learning algorithm, which utilizes knowledge learned from high-resource domains, to boost the performance of low-resource UNMT. Our model surpasses a transfer learning-based approach by up to 2-4 BLEU scores. Extensive experimental results show that our proposed algorithm is pertinent for fast adaptation and consistently outperforms other baseline models.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Cheonbok Park (20 papers)
  2. Yunwon Tae (5 papers)
  3. Taehee Kim (25 papers)
  4. Soyoung Yang (9 papers)
  5. Mohammad Azam Khan (5 papers)
  6. Eunjeong Park (3 papers)
  7. Jaegul Choo (161 papers)
Citations (16)