Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
167 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Diverse Embedding Neural Network Language Models (1412.7063v5)

Published 22 Dec 2014 in cs.CL, cs.LG, and cs.NE

Abstract: We propose Diverse Embedding Neural Network (DENN), a novel architecture for LLMs (LMs). A DENNLM projects the input word history vector onto multiple diverse low-dimensional sub-spaces instead of a single higher-dimensional sub-space as in conventional feed-forward neural network LMs. We encourage these sub-spaces to be diverse during network training through an augmented loss function. Our LLMing experiments on the Penn Treebank data set show the performance benefit of using a DENNLM.

Citations (1)

Summary

We haven't generated a summary for this paper yet.