Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Document-level Neural Machine Translation with Document Embeddings (2009.08775v1)

Published 16 Sep 2020 in cs.CL

Abstract: Standard neural machine translation (NMT) is on the assumption of document-level context independent. Most existing document-level NMT methods are satisfied with a smattering sense of brief document-level information, while this work focuses on exploiting detailed document-level context in terms of multiple forms of document embeddings, which is capable of sufficiently modeling deeper and richer document-level context. The proposed document-aware NMT is implemented to enhance the Transformer baseline by introducing both global and local document-level clues on the source end. Experiments show that the proposed method significantly improves the translation performance over strong baselines and other related studies.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Shu Jiang (18 papers)
  2. Hai Zhao (227 papers)
  3. Zuchao Li (76 papers)
  4. Bao-Liang Lu (26 papers)
Citations (2)