Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Training Multilingual Machine Translation by Alternately Freezing Language-Specific Encoders-Decoders (2006.01594v1)

Published 29 May 2020 in cs.CL

Abstract: We propose a modular architecture of language-specific encoder-decoders that constitutes a multilingual machine translation system that can be incrementally extended to new languages without the need for retraining the existing system when adding new languages. Differently from previous works, we simultaneously train $N$ languages in all translation directions by alternately freezing encoder or decoder modules, which indirectly forces the system to train in a common intermediate representation for all languages. Experimental results from multilingual machine translation show that we can successfully train this modular architecture improving on the initial languages while falling slightly behind when adding new languages or doing zero-shot translation. Additional comparison of the quality of sentence representation in the task of natural language inference shows that the alternately freezing training is also beneficial in this direction.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Carlos Escolano (20 papers)
  2. Marta R. Costa-jussà (73 papers)
  3. José A. R. Fonollosa (23 papers)
  4. Mikel Artetxe (52 papers)
Citations (10)

Summary

We haven't generated a summary for this paper yet.