Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A Shared Attention Mechanism for Interpretation of Neural Automatic Post-Editing Systems (1807.00248v1)

Published 1 Jul 2018 in cs.CL

Abstract: Automatic post-editing (APE) systems aim to correct the systematic errors made by machine translators. In this paper, we propose a neural APE system that encodes the source (src) and machine translated (mt) sentences with two separate encoders, but leverages a shared attention mechanism to better understand how the two inputs contribute to the generation of the post-edited (pe) sentences. Our empirical observations have showed that when the mt is incorrect, the attention shifts weight toward tokens in the src sentence to properly edit the incorrect translation. The model has been trained and evaluated on the official data from the WMT16 and WMT17 APE IT domain English-German shared tasks. Additionally, we have used the extra 500K artificial data provided by the shared task. Our system has been able to reproduce the accuracies of systems trained with the same data, while at the same time providing better interpretability.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Inigo Jauregi Unanue (13 papers)
  2. Ehsan Zare Borzeshi (7 papers)
  3. Massimo Piccardi (21 papers)
Citations (6)

Summary

We haven't generated a summary for this paper yet.