Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Non-Parametric Unsupervised Domain Adaptation for Neural Machine Translation (2109.06604v2)

Published 14 Sep 2021 in cs.CL, cs.AI, and cs.LG

Abstract: Recently, $k$NN-MT has shown the promising capability of directly incorporating the pre-trained neural machine translation (NMT) model with domain-specific token-level $k$-nearest-neighbor ($k$NN) retrieval to achieve domain adaptation without retraining. Despite being conceptually attractive, it heavily relies on high-quality in-domain parallel corpora, limiting its capability on unsupervised domain adaptation, where in-domain parallel corpora are scarce or nonexistent. In this paper, we propose a novel framework that directly uses in-domain monolingual sentences in the target language to construct an effective datastore for $k$-nearest-neighbor retrieval. To this end, we first introduce an autoencoder task based on the target language, and then insert lightweight adapters into the original NMT model to map the token-level representation of this task to the ideal representation of translation task. Experiments on multi-domain datasets demonstrate that our proposed approach significantly improves the translation accuracy with target-side monolingual data, while achieving comparable performance with back-translation.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Xin Zheng (57 papers)
  2. Zhirui Zhang (46 papers)
  3. Shujian Huang (106 papers)
  4. Boxing Chen (67 papers)
  5. Jun Xie (66 papers)
  6. Weihua Luo (63 papers)
  7. Jiajun Chen (125 papers)
Citations (25)