Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A Joint Learning Approach for Semi-supervised Neural Topic Modeling (2204.03208v1)

Published 7 Apr 2022 in cs.IR, cs.CL, cs.LG, and stat.ML

Abstract: Topic models are some of the most popular ways to represent textual data in an interpret-able manner. Recently, advances in deep generative models, specifically auto-encoding variational Bayes (AEVB), have led to the introduction of unsupervised neural topic models, which leverage deep generative models as opposed to traditional statistics-based topic models. We extend upon these neural topic models by introducing the Label-Indexed Neural Topic Model (LI-NTM), which is, to the extent of our knowledge, the first effective upstream semi-supervised neural topic model. We find that LI-NTM outperforms existing neural topic models in document reconstruction benchmarks, with the most notable results in low labeled data regimes and for data-sets with informative labels; furthermore, our jointly learned classifier outperforms baseline classifiers in ablation studies.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Jeffrey Chiu (1 paper)
  2. Rajat Mittal (42 papers)
  3. Neehal Tumma (3 papers)
  4. Abhishek Sharma (112 papers)
  5. Finale Doshi-Velez (134 papers)
Citations (4)

Summary

We haven't generated a summary for this paper yet.