Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Doubly-Trained Adversarial Data Augmentation for Neural Machine Translation (2110.05691v1)

Published 12 Oct 2021 in cs.CL

Abstract: Neural Machine Translation (NMT) models are known to suffer from noisy inputs. To make models robust, we generate adversarial augmentation samples that attack the model and preserve the source-side semantic meaning at the same time. To generate such samples, we propose a doubly-trained architecture that pairs two NMT models of opposite translation directions with a joint loss function, which combines the target-side attack and the source-side semantic similarity constraint. The results from our experiments across three different language pairs and two evaluation metrics show that these adversarial samples improve the model robustness.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Weiting Tan (14 papers)
  2. Shuoyang Ding (17 papers)
  3. Huda Khayrallah (15 papers)
  4. Philipp Koehn (60 papers)
Citations (1)