Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Lexical Translation Model Using a Deep Neural Network Architecture (1504.07395v1)

Published 28 Apr 2015 in cs.CL, cs.LG, and cs.NE

Abstract: In this paper we combine the advantages of a model using global source sentence contexts, the Discriminative Word Lexicon, and neural networks. By using deep neural networks instead of the linear maximum entropy model in the Discriminative Word Lexicon models, we are able to leverage dependencies between different source words due to the non-linearity. Furthermore, the models for different target words can share parameters and therefore data sparsity problems are effectively reduced. By using this approach in a state-of-the-art translation system, we can improve the performance by up to 0.5 BLEU points for three different language pairs on the TED translation task.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Thanh-Le Ha (13 papers)
  2. Jan Niehues (76 papers)
  3. Alex Waibel (48 papers)
Citations (10)