Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Specializing Multi-domain NMT via Penalizing Low Mutual Information (2210.12910v1)

Published 24 Oct 2022 in cs.CL and cs.AI

Abstract: Multi-domain Neural Machine Translation (NMT) trains a single model with multiple domains. It is appealing because of its efficacy in handling multiple domains within one model. An ideal multi-domain NMT should learn distinctive domain characteristics simultaneously, however, grasping the domain peculiarity is a non-trivial task. In this paper, we investigate domain-specific information through the lens of mutual information (MI) and propose a new objective that penalizes low MI to become higher. Our method achieved the state-of-the-art performance among the current competitive multi-domain NMT models. Also, we empirically show our objective promotes low MI to be higher resulting in domain-specialized multi-domain NMT.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Jiyoung Lee (42 papers)
  2. Hantae Kim (3 papers)
  3. Hyunchang Cho (4 papers)
  4. Edward Choi (90 papers)
  5. Cheonbok Park (20 papers)
Citations (4)

Summary

We haven't generated a summary for this paper yet.