Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

SenGen: Sentence Generating Neural Variational Topic Model (1708.00308v1)

Published 1 Aug 2017 in cs.CL, cs.LG, and stat.ML

Abstract: We present a new topic model that generates documents by sampling a topic for one whole sentence at a time, and generating the words in the sentence using an RNN decoder that is conditioned on the topic of the sentence. We argue that this novel formalism will help us not only visualize and model the topical discourse structure in a document better, but also potentially lead to more interpretable topics since we can now illustrate topics by sampling representative sentences instead of bag of words or phrases. We present a variational auto-encoder approach for learning in which we use a factorized variational encoder that independently models the posterior over topical mixture vectors of documents using a feed-forward network, and the posterior over topic assignments to sentences using an RNN. Our preliminary experiments on two different datasets indicate early promise, but also expose many challenges that remain to be addressed.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Ramesh Nallapati (38 papers)
  2. Igor Melnyk (28 papers)
  3. Abhishek Kumar (172 papers)
  4. Bowen Zhou (141 papers)
Citations (12)

Summary

We haven't generated a summary for this paper yet.