2000 character limit reached
Later-stage Minimum Bayes-Risk Decoding for Neural Machine Translation (1704.03169v2)
Published 11 Apr 2017 in cs.CL
Abstract: For extended periods of time, sequence generation models rely on beam search algorithm to generate output sequence. However, the correctness of beam search degrades when the a model is over-confident about a suboptimal prediction. In this paper, we propose to perform minimum Bayes-risk (MBR) decoding for some extra steps at a later stage. In order to speed up MBR decoding, we compute the Bayes risks on GPU in batch mode. In our experiments, we found that MBR reranking works with a large beam size. Later-stage MBR decoding is shown to outperform simple MBR reranking in machine translation tasks.
- Raphael Shu (24 papers)
- Hideki Nakayama (59 papers)