Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Improving Variational Autoencoder for Text Modelling with Timestep-Wise Regularisation (2011.01136v2)

Published 2 Nov 2020 in cs.CL and cs.LG

Abstract: The Variational Autoencoder (VAE) is a popular and powerful model applied to text modelling to generate diverse sentences. However, an issue known as posterior collapse (or KL loss vanishing) happens when the VAE is used in text modelling, where the approximate posterior collapses to the prior, and the model will totally ignore the latent variables and be degraded to a plain LLM during text generation. Such an issue is particularly prevalent when RNN-based VAE models are employed for text modelling. In this paper, we propose a simple, generic architecture called Timestep-Wise Regularisation VAE (TWR-VAE), which can effectively avoid posterior collapse and can be applied to any RNN-based VAE models. The effectiveness and versatility of our model are demonstrated in different tasks, including LLMling and dialogue response generation.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Ruizhe Li (40 papers)
  2. Xiao Li (357 papers)
  3. Guanyi Chen (26 papers)
  4. Chenghua Lin (127 papers)
Citations (17)

Summary

We haven't generated a summary for this paper yet.