Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Universal Vector Neural Machine Translation With Effective Attention (2006.05003v1)

Published 9 Jun 2020 in cs.CL

Abstract: Neural Machine Translation (NMT) leverages one or more trained neural networks for the translation of phrases. Sutskever introduced a sequence to sequence based encoder-decoder model which became the standard for NMT based systems. Attention mechanisms were later introduced to address the issues with the translation of long sentences and improving overall accuracy. In this paper, we propose a singular model for Neural Machine Translation based on encoder-decoder models. Most translation models are trained as one model for one translation. We introduce a neutral/universal model representation that can be used to predict more than one language depending on the source and a provided target. Secondly, we introduce an attention model by adding an overall learning vector to the multiplicative model. With these two changes, by using the novel universal model the number of models needed for multiple language translation applications are reduced.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Satish Mylapore (1 paper)
  2. Ryan Quincy Paul (1 paper)
  3. Joshua Yi (1 paper)
  4. Robert D. Slater (1 paper)
Citations (1)

Summary

We haven't generated a summary for this paper yet.