Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Compressing Neural Language Models by Sparse Word Representations (1610.03950v1)

Published 13 Oct 2016 in cs.CL and cs.LG

Abstract: Neural networks are among the state-of-the-art techniques for LLMing. Existing neural LLMs typically map discrete words to distributed, dense vector representations. After information processing of the preceding context words by hidden layers, an output layer estimates the probability of the next word. Such approaches are time- and memory-intensive because of the large numbers of parameters for word embeddings and the output layer. In this paper, we propose to compress neural LLMs by sparse word representations. In the experiments, the number of parameters in our model increases very slowly with the growth of the vocabulary size, which is almost imperceptible. Moreover, our approach not only reduces the parameter space to a large extent, but also improves the performance in terms of the perplexity measure.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Yunchuan Chen (6 papers)
  2. Lili Mou (79 papers)
  3. Yan Xu (258 papers)
  4. Ge Li (213 papers)
  5. Zhi Jin (160 papers)
Citations (29)

Summary

We haven't generated a summary for this paper yet.