Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Syntax-Aware Language Modeling with Recurrent Neural Networks (1803.03665v1)

Published 2 Mar 2018 in cs.CL and cs.LG

Abstract: Neural LLMs (LMs) are typically trained using only lexical features, such as surface forms of words. In this paper, we argue this deprives the LM of crucial syntactic signals that can be detected at high confidence using existing parsers. We present a simple but highly effective approach for training neural LMs using both lexical and syntactic information, and a novel approach for applying such LMs to unparsed text using sequential Monte Carlo sampling. In experiments on a range of corpora and corpus sizes, we show our approach consistently outperforms standard lexical LMs in character-level LLMing; on the other hand, for word-level models the models are on a par with standard LLMs. These results indicate potential for expanding LMs beyond lexical surface features to higher-level NLP features for character-level models.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Duncan Blythe (3 papers)
  2. Alan Akbik (26 papers)
  3. Roland Vollgraf (17 papers)
Citations (2)

Summary

We haven't generated a summary for this paper yet.