Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Analyzing Chain-of-Thought Prompting in Large Language Models via Gradient-based Feature Attributions (2307.13339v1)

Published 25 Jul 2023 in cs.CL and cs.AI

Abstract: Chain-of-thought (CoT) prompting has been shown to empirically improve the accuracy of LLMs on various question answering tasks. While understanding why CoT prompting is effective is crucial to ensuring that this phenomenon is a consequence of desired model behavior, little work has addressed this; nonetheless, such an understanding is a critical prerequisite for responsible model deployment. We address this question by leveraging gradient-based feature attribution methods which produce saliency scores that capture the influence of input tokens on model output. Specifically, we probe several open-source LLMs to investigate whether CoT prompting affects the relative importances they assign to particular input tokens. Our results indicate that while CoT prompting does not increase the magnitude of saliency scores attributed to semantically relevant tokens in the prompt compared to standard few-shot prompting, it increases the robustness of saliency scores to question perturbations and variations in model output.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Skyler Wu (5 papers)
  2. Eric Meng Shen (1 paper)
  3. Charumathi Badrinath (4 papers)
  4. Jiaqi Ma (82 papers)
  5. Himabindu Lakkaraju (88 papers)
Citations (21)