Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
112 tokens/sec
GPT-4o
12 tokens/sec
Gemini 2.5 Pro Pro
39 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
5 tokens/sec
DeepSeek R1 via Azure Pro
33 tokens/sec
2000 character limit reached

Large Language Models and the Extended Church-Turing Thesis (2409.06978v1)

Published 11 Sep 2024 in cs.FL and cs.AI

Abstract: The Extended Church-Turing Thesis (ECTT) posits that all effective information processing, including unbounded and non-uniform interactive computations, can be described in terms of interactive Turing machines with advice. Does this assertion also apply to the abilities of contemporary LLMs? From a broader perspective, this question calls for an investigation of the computational power of LLMs by the classical means of computability and computational complexity theory, especially the theory of automata. Along these lines, we establish a number of fundamental results. Firstly, we argue that any fixed (non-adaptive) LLM is computationally equivalent to a, possibly very large, deterministic finite-state transducer. This characterizes the base level of LLMs. We extend this to a key result concerning the simulation of space-bounded Turing machines by LLMs. Secondly, we show that lineages of evolving LLMs are computationally equivalent to interactive Turing machines with advice. The latter finding confirms the validity of the ECTT for lineages of LLMs. From a computability viewpoint, it also suggests that lineages of LLMs possess super-Turing computational power. Consequently, in our computational model knowledge generation is in general a non-algorithmic process realized by lineages of LLMs. Finally, we discuss the merits of our findings in the broader context of several related disciplines and philosophies.

Summary

  • The paper demonstrates that fixed LLMs are computationally equivalent to deterministic finite-state transducers through mutual simulation, reinforcing the ECTT framework.
  • The paper shows that LLMs can simulate space-bounded Turing machines using embeddings proportional to input complexity, highlighting their resource constraints.
  • The paper introduces evolving LLM lineages that match interactive Turing machines with advice, paving the way for adaptive AI systems within classical computation limits.

LLMs and the Extended Church-Turing Thesis

The paper "LLMs and the Extended Church-Turing Thesis" by Jiří Wiedermann and Jan van Leeuwen examines the applicability of the Extended Church-Turing Thesis (ECTT) to contemporary LLMs. The ECTT posits that all effective forms of information processing, including unbounded and non-uniform interactive computations, can be modeled using interactive Turing machines with advice (ITM/As). The primary investigation of this paper is whether LLMs fall within the computational boundaries defined by the ECTT.

Introduction

The authors articulate the historical context of the ECTT, which extends the traditional Church-Turing Thesis by incorporating developments in modern computing—a paradigm characterized by non-uniformity, interactivity, and potentially infinite operations. They compare two seemingly disparate models of computation: interactive Turing machines with advice and LLMs. The applicability of the ECTT to LLMs is significant, as it supports the hypothesis that these models' computational power remains within the confines of interactive Turing machine limits.

Computational Power of LLMs

Equivalence to Finite-State Transducers

The paper initially posits that any fixed (non-adaptive) LLM is computationally equivalent to a deterministic finite-state transducer (FST). This equivalence is demonstrated through mutual simulation:

  1. Simulation of LLMs by FSTs:
    • The authors argue that fixed LLMs, being deterministic and of bounded size, can be simulated by a very large deterministic FST.
  2. Simulation of FSTs by LLMs:
    • For a deterministic FST, there exists a fixed LLM that can simulate the FST for inputs of any fixed length, n. The simulation relies on the learned behavior of LLMs, leveraging their vector embeddings to mimic FST transitions.

This leads to the fundamental conclusion that the computational power of fixed LLMs equals that of deterministic FSTs.

Simulation of Space-Bounded Turing Machines

Taking a step further, the authors demonstrate that LLMs can simulate space-bounded Turing machines (TMs). Given a TM of space complexity S(n), there exists an LLM with word-to-vector embeddings of size proportional to S(n) that can simulate the TM on any input of length n. This insight suggests that while LLMs can internalize complex computational processes, their fixed nature confines them within specific resource limits.

Lineages of Evolving LLMs

A more powerful concept introduced is that of lineages of evolving LLMs. These are sequences of LLMs where each member can progressively handle larger computational complexities and adjust based on new inputs and advice. The authors argue that such lineages are computationally equivalent to ITM/As, thus effectively validating the ECTT for LLMs. This equivalence indicates that lineages of LLMs possess super-Turing computational power—surpassing classical Turing machines but remaining within the field defined by ECTT.

Implications and Future Directions

The implications of these findings are manifold:

  1. Practical Implications:
    • The simulations signify that LLMs, despite their finite-state nature, can effectively perform a wide range of computations typically reserved for more complex machine models. This positions LLMs as robust tools for diverse AI applications while delineating their inherent computational boundaries.
  2. Theoretical Impact:
    • By demonstrating the equivalence between LLMs and fundamental automata theory constructs, the research bridges classical computation models with modern AI architectures. This synthesis promotes deeper investigations into the nature of non-uniform, interactive, and resource-bounded computations.
  3. Cognitive and Philosophical Insights:
    • The author's analogy between natural language processing by LLMs and TM computations provides profound insights into AI's capabilities and limitations in understanding and generating human-like responses. These insights stimulate discussions on the essence of AI in replicating human cognition and language comprehension.

Conclusion

The research underscores the relevance of the ECTT in the context of modern computing systems like LLMs. It validates that LLMs fall within the scope of interactive Turing machines with advice, reinforcing their computational viability within established theoretical models. Additionally, the paper suggests that the continued evolution and scaling of LLMs will likely adhere to these computational frameworks, paving the way for more advanced, adaptive, and interactive AI systems.

The theoretical rigor and practical implications presented invite further exploration and validation, particularly in contextualizing AI's ability to simulate more complex, dynamic computational models while adhering to computability theory paradigms.

Dice Question Streamline Icon: https://streamlinehq.com

Follow-up Questions

We haven't generated follow-up questions for this paper yet.