Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

On the Representational Capacity of Recurrent Neural Language Models (2310.12942v5)

Published 19 Oct 2023 in cs.CL and cs.LG

Abstract: This work investigates the computational expressivity of LLMs (LMs) based on recurrent neural networks (RNNs). Siegelmann and Sontag (1992) famously showed that RNNs with rational weights and hidden states and unbounded computation time are Turing complete. However, LMs define weightings over strings in addition to just (unweighted) language membership and the analysis of the computational power of RNN LMs (RLMs) should reflect this. We extend the Turing completeness result to the probabilistic case, showing how a rationally weighted RLM with unbounded computation time can simulate any deterministic probabilistic Turing machine (PTM) with rationally weighted transitions. Since, in practice, RLMs work in real-time, processing a symbol at every time step, we treat the above result as an upper bound on the expressivity of RLMs. We also provide a lower bound by showing that under the restriction to real-time computation, such models can simulate deterministic real-time rational PTMs.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Franz Nowak (8 papers)
  2. Anej Svete (20 papers)
  3. Li Du (72 papers)
  4. Ryan Cotterell (226 papers)
Citations (1)

Summary

We haven't generated a summary for this paper yet.