Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Learning to Copy for Automatic Post-Editing (1911.03627v1)

Published 9 Nov 2019 in cs.CL

Abstract: Automatic post-editing (APE), which aims to correct errors in the output of machine translation systems in a post-processing step, is an important task in natural language processing. While recent work has achieved considerable performance gains by using neural networks, how to model the copying mechanism for APE remains a challenge. In this work, we propose a new method for modeling copying for APE. To better identify translation errors, our method learns the representations of source sentences and system outputs in an interactive way. These representations are used to explicitly indicate which words in the system outputs should be copied, which is useful to help CopyNet (Gu et al., 2016) better generate post-edited translations. Experiments on the datasets of the WMT 2016-2017 APE shared tasks show that our approach outperforms all best published results.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Xuancheng Huang (7 papers)
  2. Yang Liu (2253 papers)
  3. Huanbo Luan (15 papers)
  4. Jingfang Xu (11 papers)
  5. Maosong Sun (337 papers)
Citations (10)

Summary

We haven't generated a summary for this paper yet.