Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Improve Variational Autoencoder for Text Generationwith Discrete Latent Bottleneck (2004.10603v2)

Published 22 Apr 2020 in cs.LG, cs.CL, and stat.ML

Abstract: Variational autoencoders (VAEs) are essential tools in end-to-end representation learning. However, the sequential text generation common pitfall with VAEs is that the model tends to ignore latent variables with a strong auto-regressive decoder. In this paper, we propose a principled approach to alleviate this issue by applying a discretized bottleneck to enforce an implicit latent feature matching in a more compact latent space. We impose a shared discrete latent space where each input is learned to choose a combination of latent atoms as a regularized latent representation. Our model endows a promising capability to model underlying semantics of discrete sequences and thus provide more interpretative latent structures. Empirically, we demonstrate our model's efficiency and effectiveness on a broad range of tasks, including LLMing, unaligned text style transfer, dialog response generation, and neural machine translation.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Yang Zhao (382 papers)
  2. Ping Yu (42 papers)
  3. Suchismit Mahapatra (6 papers)
  4. Qinliang Su (30 papers)
  5. Changyou Chen (108 papers)
Citations (1)