Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Dirichlet Variational Autoencoder for Text Modeling (1811.00135v1)

Published 31 Oct 2018 in cs.CL, cs.AI, cs.LG, and cs.NE

Abstract: We introduce an improved variational autoencoder (VAE) for text modeling with topic information explicitly modeled as a Dirichlet latent variable. By providing the proposed model topic awareness, it is more superior at reconstructing input texts. Furthermore, due to the inherent interactions between the newly introduced Dirichlet variable and the conventional multivariate Gaussian variable, the model is less prone to KL divergence vanishing. We derive the variational lower bound for the new model and conduct experiments on four different data sets. The results show that the proposed model is superior at text reconstruction across the latent space and classifications on learned representations have higher test accuracies.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Yijun Xiao (10 papers)
  2. Tiancheng Zhao (48 papers)
  3. William Yang Wang (254 papers)
Citations (21)