Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

How Do Source-side Monolingual Word Embeddings Impact Neural Machine Translation? (1806.01515v2)

Published 5 Jun 2018 in cs.CL

Abstract: Using pre-trained word embeddings as input layer is a common practice in many NLP tasks, but it is largely neglected for neural machine translation (NMT). In this paper, we conducted a systematic analysis on the effect of using pre-trained source-side monolingual word embedding in NMT. We compared several strategies, such as fixing or updating the embeddings during NMT training on varying amounts of data, and we also proposed a novel strategy called dual-embedding that blends the fixing and updating strategies. Our results suggest that pre-trained embeddings can be helpful if properly incorporated into NMT, especially when parallel data is limited or additional in-domain monolingual data is readily available.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (2)
  1. Shuoyang Ding (17 papers)
  2. Kevin Duh (65 papers)
Citations (4)

Summary

We haven't generated a summary for this paper yet.