Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Lower Bounds on the Expressivity of Recurrent Neural Language Models (2405.19222v2)

Published 29 May 2024 in cs.CL

Abstract: The recent successes and spread of large neural LLMs (LMs) call for a thorough understanding of their computational ability. Describing their computational abilities through LMs' \emph{representational capacity} is a lively area of research. However, investigation into the representational capacity of neural LMs has predominantly focused on their ability to \emph{recognize} formal languages. For example, recurrent neural networks (RNNs) with Heaviside activations are tightly linked to regular languages, i.e., languages defined by finite-state automata (FSAs). Such results, however, fall short of describing the capabilities of RNN \emph{LLMs} (LMs), which are definitionally \emph{distributions} over strings. We take a fresh look at the representational capacity of RNN LMs by connecting them to \emph{probabilistic} FSAs and demonstrate that RNN LMs with linearly bounded precision can express arbitrary regular LMs.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Anej Svete (20 papers)
  2. Franz Nowak (8 papers)
  3. Anisha Mohamed Sahabdeen (1 paper)
  4. Ryan Cotterell (226 papers)
X Twitter Logo Streamline Icon: https://streamlinehq.com