Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Variational Recurrent Neural Machine Translation (1801.05119v1)

Published 16 Jan 2018 in cs.CL

Abstract: Partially inspired by successful applications of variational recurrent neural networks, we propose a novel variational recurrent neural machine translation (VRNMT) model in this paper. Different from the variational NMT, VRNMT introduces a series of latent random variables to model the translation procedure of a sentence in a generative way, instead of a single latent variable. Specifically, the latent random variables are included into the hidden states of the NMT decoder with elements from the variational autoencoder. In this way, these variables are recurrently generated, which enables them to further capture strong and complex dependencies among the output translations at different timesteps. In order to deal with the challenges in performing efficient posterior inference and large-scale training during the incorporation of latent variables, we build a neural posterior approximator, and equip it with a reparameterization technique to estimate the variational lower bound. Experiments on Chinese-English and English-German translation tasks demonstrate that the proposed model achieves significant improvements over both the conventional and variational NMT models.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Jinsong Su (96 papers)
  2. Shan Wu (29 papers)
  3. Deyi Xiong (103 papers)
  4. Yaojie Lu (61 papers)
  5. Xianpei Han (103 papers)
  6. Biao Zhang (76 papers)
Citations (83)

Summary

We haven't generated a summary for this paper yet.