Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Topic Compositional Neural Language Model (1712.09783v3)

Published 28 Dec 2017 in cs.LG and cs.CL

Abstract: We propose a Topic Compositional Neural LLM (TCNLM), a novel method designed to simultaneously capture both the global semantic meaning and the local word ordering structure in a document. The TCNLM learns the global semantic coherence of a document via a neural topic model, and the probability of each learned latent topic is further used to build a Mixture-of-Experts (MoE) LLM, where each expert (corresponding to one topic) is a recurrent neural network (RNN) that accounts for learning the local structure of a word sequence. In order to train the MoE model efficiently, a matrix factorization method is applied, by extending each weight matrix of the RNN to be an ensemble of topic-dependent weight matrices. The degree to which each member of the ensemble is used is tied to the document-dependent probability of the corresponding topics. Experimental results on several corpora show that the proposed approach outperforms both a pure RNN-based model and other topic-guided LLMs. Further, our model yields sensible topics, and also has the capacity to generate meaningful sentences conditioned on given topics.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Wenlin Wang (27 papers)
  2. Zhe Gan (135 papers)
  3. Wenqi Wang (29 papers)
  4. Dinghan Shen (34 papers)
  5. Jiaji Huang (17 papers)
  6. Wei Ping (51 papers)
  7. Sanjeev Satheesh (14 papers)
  8. Lawrence Carin (203 papers)
Citations (79)