Necessity of explicit configuration dictionary in LLM-based Turing machine simulations
Determine whether simulating deterministic multi-tape Turing machines and Turing machines with advice entirely within large language models by representing tape configurations as word-to-vector embeddings necessarily requires an explicit representation of the full dictionary of all required Turing-machine-configuration “words,” or whether alternative internal mechanisms of large language models can achieve the simulation without such explicit dictionary storage, and specify the architectural assumptions under which such improvements would be possible.
Sponsor
References
It is an open problem whether the simulations from Theorem \ref{thm:interior} and Corollary \ref{cor:interior} can be improved. For instance, does the model need to explicitly represent the full dictionary of the necessary ``Turing machine words"? It seems to depend on the possibilities of the internal model of an LLM.