Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Chain-of-Probe: Examing the Necessity and Accuracy of CoT Step-by-Step (2406.16144v1)

Published 23 Jun 2024 in cs.CL

Abstract: Current research found the issue of Early Answering in LLMs, where the models already have an answer before generating the Chain-of-Thought (CoT). This phenomenon suggests a potential lack of necessary dependency between the predicted answer and the reasoning process. Consequently, two important questions arise: (1) Is CoT still necessary if the model already has an answer? (2) Can the correctness of the answer serve as valid evidence for the correctness of CoT? To address these questions, we propose a method, namely Chain-of-Probe (CoP), to probe changes in the mind during the model's reasoning. The probing results show that in a significant number of question-answer cases, CoT appears to be unnecessary, and this necessity correlates with the simplicity of the task, defined by reasoning steps required. Furthermore, by analyzing patterns in mind change, we examine the correctness of the model's reasoning. Our validation reveals that many responses, although correct in their final answer, contain errors in their reasoning process. To this end, we propose a strategic approach based on CoP to prioritize answers with correct reasoning among multiple candidates, thereby bolstering the reliability of the model's reasoning.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (10)
  1. Zezhong Wang (30 papers)
  2. Xingshan Zeng (38 papers)
  3. Weiwen Liu (59 papers)
  4. Yufei Wang (141 papers)
  5. Liangyou Li (36 papers)
  6. Yasheng Wang (91 papers)
  7. Lifeng Shang (90 papers)
  8. Xin Jiang (242 papers)
  9. Qun Liu (230 papers)
  10. Kam-Fai Wong (92 papers)