Let's Think Dot by Dot: Hidden Computation in Transformer Language Models (2404.15758v1)
Abstract: Chain-of-thought responses from LLMs improve performance across most benchmarks. However, it remains unclear to what extent these performance gains can be attributed to human-like task decomposition or simply the greater computation that additional tokens allow. We show that transformers can use meaningless filler tokens (e.g., '......') in place of a chain of thought to solve two hard algorithmic tasks they could not solve when responding without intermediate tokens. However, we find empirically that learning to use filler tokens is difficult and requires specific, dense supervision to converge. We also provide a theoretical characterization of the class of problems where filler tokens are useful in terms of the quantifier depth of a first-order formula. For problems satisfying this characterization, chain-of-thought tokens need not provide information about the intermediate computational steps involved in multi-token computations. In summary, our results show that additional tokens can provide computational benefits independent of token choice. The fact that intermediate tokens can act as filler tokens raises concerns about LLMs engaging in unauditable, hidden computations that are increasingly detached from the observed chain-of-thought tokens.
- Towards revealing the mystery behind chain of thought: A theoretical perspective. In Thirty-seventh Conference on Neural Information Processing Systems, 2023. URL https://openreview.net/forum?id=qHrADgAdYu.
- Think before you speak: Training language models with pause tokens. 2024. URL https://arxiv.org/abs/2310.02226.
- Why are sensitive functions hard for transformers? 2024. URL https://arxiv.org/abs/2402.09963.
- Designing and Interpreting Probes with Control Tasks. In Kentaro Inui, Jing Jiang, Vincent Ng, and Xiaojun Wan (eds.), Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pp. 2733–2743, Hong Kong, China, November 2019. Association for Computational Linguistics. doi: 10.18653/v1/D19-1275. URL https://aclanthology.org/D19-1275.
- Janus. How LLMs are and are not myopic. 2023. URL https://www.lesswrong.com/posts/c68SJsBpiAxkPwRHj/how-llms-are-and-are-not-myopic.
- Measuring faithfulness in chain-of-thought reasoning. arXiv preprint arXiv:2307.13702, 2023. URL https://arxiv.org/abs/2307.13702.
- The parallelism tradeoff: Limitations of log-precision transformers. Transactions of the Association for Computational Linguistics, 11:531–545, 2023a. doi: 10.1162/tacl˙a˙00562. URL https://aclanthology.org/2023.tacl-1.31.
- A logic for expressing log-precision transformers. In Thirty-seventh Conference on Neural Information Processing Systems, 2023b. URL https://openreview.net/forum?id=uR8TtWCIsr.
- The expressive power of transformers with chain of thought, 2023c. URL https://arxiv.org/abs/2310.07923.
- Future Lens: Anticipating Subsequent Tokens from a Single Hidden State. In Proceedings of the 27th Conference on Computational Natural Language Learning (CoNLL), pp. 548–560, 2023. doi: 10.18653/v1/2023.conll-1.37. URL http://arxiv.org/abs/2311.04897. arXiv:2311.04897 [cs].
- Kshitij Sachan. Llms are (mostly) not helped by filler tokens, 2023. URL https://www.lesswrong.com/posts/oSZ2xTxEMZh9f3Yaz/llms-are-mostly-not-helped-by-filler-tokens.
- Representational strengths and limitations of transformers. Advances in Neural Information Processing Systems, 36, 2024. URL https://arxiv.org/abs/2306.02896.
- Transformers as recognizers of formal languages: A survey on expressivity, 2023. URL https://arxiv.org/abs/2311.00208.
- Challenging big-bench tasks and whether chain-of-thought can solve them. 2022. URL https://arxiv.org/abs/2210.09261.
- Llama: Open and efficient foundation language models, 2023. URL https://arxiv.org/abs/2302.13971.
- Language models don't always say what they think: Unfaithful explanations in chain-of-thought prompting. In A. Oh, T. Neumann, A. Globerson, K. Saenko, M. Hardt, and S. Levine (eds.), Advances in Neural Information Processing Systems, volume 36, pp. 74952–74965. Curran Associates, Inc., 2023. URL https://proceedings.neurips.cc/paper_files/paper/2023/file/ed3fea9033a80fea1376299fa7863f4a-Paper-Conference.pdf.
- Chain-of-thought prompting elicits reasoning in large language models. 2023. URL https://arxiv.org/abs/2201.11903.
- Do language models plan ahead for future tokens?, March 2024. URL http://arxiv.org/abs/2404.00859. arXiv:2404.00859 [cs].
- Quiet-star: Language models can teach themselves to think before speaking. 2024. URL https://arxiv.org/abs/2403.09629.