Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Character-Word LSTM Language Models (1704.02813v1)

Published 10 Apr 2017 in cs.CL

Abstract: We present a Character-Word Long Short-Term Memory LLM which both reduces the perplexity with respect to a baseline word-level LLM and reduces the number of parameters of the model. Character information can reveal structural (dis)similarities between words and can even be used when a word is out-of-vocabulary, thus improving the modeling of infrequent and unknown words. By concatenating word and character embeddings, we achieve up to 2.77% relative improvement on English compared to a baseline model with a similar amount of parameters and 4.57% on Dutch. Moreover, we also outperform baseline word-level models with a larger number of parameters.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Lyan Verwimp (11 papers)
  2. Joris Pelemans (7 papers)
  3. Hugo Van hamme (59 papers)
  4. Patrick Wambacq (5 papers)
Citations (51)

Summary

We haven't generated a summary for this paper yet.