Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Pivot-based Transfer Learning for Neural Machine Translation between Non-English Languages (1909.09524v1)

Published 20 Sep 2019 in cs.CL and cs.LG

Abstract: We present effective pre-training strategies for neural machine translation (NMT) using parallel corpora involving a pivot language, i.e., source-pivot and pivot-target, leading to a significant improvement in source-target translation. We propose three methods to increase the relation among source, pivot, and target languages in the pre-training: 1) step-wise training of a single model for different language pairs, 2) additional adapter component to smoothly connect pre-trained encoder and decoder, and 3) cross-lingual encoder training via autoencoding of the pivot language. Our methods greatly outperform multilingual models up to +2.6% BLEU in WMT 2019 French-German and German-Czech tasks. We show that our improvements are valid also in zero-shot/zero-resource scenarios.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Yunsu Kim (40 papers)
  2. Petre Petrov (1 paper)
  3. Pavel Petrushkov (9 papers)
  4. Shahram Khadivi (29 papers)
  5. Hermann Ney (104 papers)
Citations (75)

Summary

We haven't generated a summary for this paper yet.