2000 character limit reached
Synthetic Source Language Augmentation for Colloquial Neural Machine Translation (2012.15178v1)
Published 30 Dec 2020 in cs.CL and cs.LG
Abstract: Neural machine translation (NMT) is typically domain-dependent and style-dependent, and it requires lots of training data. State-of-the-art NMT models often fall short in handling colloquial variations of its source language and the lack of parallel data in this regard is a challenging hurdle in systematically improving the existing models. In this work, we develop a novel colloquial Indonesian-English test-set collected from YouTube transcript and Twitter. We perform synthetic style augmentation to the source of formal Indonesian language and show that it improves the baseline Id-En models (in BLEU) over the new test data.
- Asrul Sani Ariesandy (1 paper)
- Mukhlis Amien (4 papers)
- Alham Fikri Aji (94 papers)
- Radityo Eko Prasojo (13 papers)