Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Word-based Domain Adaptation for Neural Machine Translation (1906.03129v1)

Published 7 Jun 2019 in cs.CL and cs.AI

Abstract: In this paper, we empirically investigate applying word-level weights to adapt neural machine translation to e-commerce domains, where small e-commerce datasets and large out-of-domain datasets are available. In order to mine in-domain like words in the out-of-domain datasets, we compute word weights by using a domain-specific and a non-domain-specific LLM followed by smoothing and binary quantization. The baseline model is trained on mixed in-domain and out-of-domain datasets. Experimental results on English to Chinese e-commerce domain translation show that compared to continuing training without word weights, it improves MT quality by up to 2.11% BLEU absolute and 1.59% TER. We have also trained models using fine-tuning on the in-domain data. Pre-training a model with word weights improves fine-tuning up to 1.24% BLEU absolute and 1.64% TER, respectively.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Shen Yan (47 papers)
  2. Leonard Dahlmann (5 papers)
  3. Pavel Petrushkov (9 papers)
  4. Sanjika Hewavitharana (5 papers)
  5. Shahram Khadivi (29 papers)
Citations (7)

Summary

We haven't generated a summary for this paper yet.