Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Towards Bidirectional Hierarchical Representations for Attention-Based Neural Machine Translation (1707.05114v1)

Published 17 Jul 2017 in cs.CL

Abstract: This paper proposes a hierarchical attentional neural translation model which focuses on enhancing source-side hierarchical representations by covering both local and global semantic information using a bidirectional tree-based encoder. To maximize the predictive likelihood of target words, a weighted variant of an attention mechanism is used to balance the attentive information between lexical and phrase vectors. Using a tree-based rare word encoding, the proposed model is extended to sub-word level to alleviate the out-of-vocabulary (OOV) problem. Empirical results reveal that the proposed model significantly outperforms sequence-to-sequence attention-based and tree-based neural translation models in English-Chinese translation tasks.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Baosong Yang (57 papers)
  2. Derek F. Wong (69 papers)
  3. Tong Xiao (119 papers)
  4. Lidia S. Chao (41 papers)
  5. Jingbo Zhu (79 papers)
Citations (33)

Summary

We haven't generated a summary for this paper yet.