2000 character limit reached
Impact of Domain-Adapted Multilingual Neural Machine Translation in the Medical Domain (2212.02143v1)
Published 5 Dec 2022 in cs.CL
Abstract: Multilingual Neural Machine Translation (MNMT) models leverage many language pairs during training to improve translation quality for low-resource languages by transferring knowledge from high-resource languages. We study the quality of a domain-adapted MNMT model in the medical domain for English-Romanian with automatic metrics and a human error typology annotation which includes terminology-specific error categories. We compare the out-of-domain MNMT with the in-domain adapted MNMT. The in-domain MNMT model outperforms the out-of-domain MNMT in all measured automatic metrics and produces fewer terminology errors.
- Miguel Rios (6 papers)
- Raluca-Maria Chereji (1 paper)
- Alina Secara (1 paper)
- Dragos Ciobanu (1 paper)