Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Saliency-driven Word Alignment Interpretation for Neural Machine Translation (1906.10282v2)

Published 25 Jun 2019 in cs.CL

Abstract: Despite their original goal to jointly learn to align and translate, Neural Machine Translation (NMT) models, especially Transformer, are often perceived as not learning interpretable word alignments. In this paper, we show that NMT models do learn interpretable word alignments, which could only be revealed with proper interpretation methods. We propose a series of such methods that are model-agnostic, are able to be applied either offline or online, and do not require parameter update or architectural change. We show that under the force decoding setup, the alignments induced by our interpretation method are of better quality than fast-align for some systems, and when performing free decoding, they agree well with the alignments induced by automatic alignment tools.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Shuoyang Ding (17 papers)
  2. Hainan Xu (20 papers)
  3. Philipp Koehn (60 papers)
Citations (55)

Summary

We haven't generated a summary for this paper yet.