Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Pivot-based Transfer Learning for Neural Machine Translation between Non-English Languages (1909.09524v1)

Published 20 Sep 2019 in cs.CL and cs.LG

Abstract: We present effective pre-training strategies for neural machine translation (NMT) using parallel corpora involving a pivot language, i.e., source-pivot and pivot-target, leading to a significant improvement in source-target translation. We propose three methods to increase the relation among source, pivot, and target languages in the pre-training: 1) step-wise training of a single model for different language pairs, 2) additional adapter component to smoothly connect pre-trained encoder and decoder, and 3) cross-lingual encoder training via autoencoding of the pivot language. Our methods greatly outperform multilingual models up to +2.6% BLEU in WMT 2019 French-German and German-Czech tasks. We show that our improvements are valid also in zero-shot/zero-resource scenarios.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Yunsu Kim (40 papers)
  2. Petre Petrov (1 paper)
  3. Pavel Petrushkov (9 papers)
  4. Shahram Khadivi (29 papers)
  5. Hermann Ney (104 papers)
Citations (75)