Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Cross-Examiner: Evaluating Consistency of Large Language Model-Generated Explanations (2503.08815v1)

Published 11 Mar 2025 in cs.CL and cs.AI

Abstract: LLMs are often asked to explain their outputs to enhance accuracy and transparency. However, evidence suggests that these explanations can misrepresent the models' true reasoning processes. One effective way to identify inaccuracies or omissions in these explanations is through consistency checking, which typically involves asking follow-up questions. This paper introduces, cross-examiner, a new method for generating follow-up questions based on a model's explanation of an initial question. Our method combines symbolic information extraction with LLM-driven question generation, resulting in better follow-up questions than those produced by LLMs alone. Additionally, this approach is more flexible than other methods and can generate a wider variety of follow-up questions.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Danielle Villa (1 paper)
  2. Maria Chang (14 papers)
  3. Keerthiram Murugesan (38 papers)
  4. Rosario Uceda-Sosa (8 papers)
  5. Karthikeyan Natesan Ramamurthy (68 papers)