Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
51 tokens/sec
GPT-4o
60 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Do Models Explain Themselves? Counterfactual Simulatability of Natural Language Explanations (2307.08678v1)

Published 17 Jul 2023 in cs.CL, cs.AI, and cs.LG
Do Models Explain Themselves? Counterfactual Simulatability of Natural Language Explanations

Abstract: LLMs are trained to imitate humans to explain human decisions. However, do LLMs explain themselves? Can they help humans build mental models of how LLMs process different inputs? To answer these questions, we propose to evaluate $\textbf{counterfactual simulatability}$ of natural language explanations: whether an explanation can enable humans to precisely infer the model's outputs on diverse counterfactuals of the explained input. For example, if a model answers "yes" to the input question "Can eagles fly?" with the explanation "all birds can fly", then humans would infer from the explanation that it would also answer "yes" to the counterfactual input "Can penguins fly?". If the explanation is precise, then the model's answer should match humans' expectations. We implemented two metrics based on counterfactual simulatability: precision and generality. We generated diverse counterfactuals automatically using LLMs. We then used these metrics to evaluate state-of-the-art LLMs (e.g., GPT-4) on two tasks: multi-hop factual reasoning and reward modeling. We found that LLM's explanations have low precision and that precision does not correlate with plausibility. Therefore, naively optimizing human approvals (e.g., RLHF) may not be a sufficient solution.

Counterfactual Simulatability: Evaluating Explanations in LLMs

The paper "Do Models Explain Themselves? Counterfactual Simulatability of Natural Language Explanations" examines the extent to which LLMs like GPT-3.5 and GPT-4 can provide explanations that help humans build mental models predicting the model's behavior across counterfactual scenarios. This paper is significant in advancing our understanding of AI explainability, particularly in contexts where human users interact with AI systems.

Methodological Framework

The authors introduce two key metrics to assess the counterfactual simulatability of LLM-generated explanations: simulation generality and simulation precision. Simulation generality refers to how diverse counterfactuals an explanation can be applied to, while simulation precision measures the consistency between the humans' inferred outputs based on the explanation and the model's actual outputs across these counterfactuals. Through automated LLM techniques, counterfactual scenarios are generated for initial model outputs and explanations.

To ascertain the efficacy of these metrics, the paper evaluates LLM-generated explanations across two tasks: multi-hop factual reasoning using StrategyQA and reward modeling with Stanford Human Preference datasets. Two explanation frameworks, Chain-of-Thought (CoT) and Post-Hoc, are juxtaposed, the former generating explanations prior to model output decision, posited as more aligned with the model's decision-making process.

Key Findings

  1. Explanatory Precision: Both CoT and Post-Hoc explanation methods show that the explanatory precision of model outputs is deficient. Particularly, the explanations provided often fail to match the inferred expectations set by humans when tasked with counterfactual simulations. Notably, the research indicates that CoT does not outperform Post-Hoc in a statistically significant manner, which is contrary to initial expectations that CoT would yield higher precision due to its process alignment.
  2. Generality versus Precision: The paper highlights a lack of correlation between simulation generality and precision, suggesting that the ability of an LLM explanation to generalize across diverse counterfactuals does not necessarily indicate higher precision in matching anticipated outputs with actual outputs.
  3. LLM Scaling and Precision: With GPT-4, explanations were observed to yield better alignment with human mental model predictions than those from GPT-3.5, indicating a potential improvement in counterfactual simulatability through scale.
  4. Dissonance with Human Preferences: A critical insight is that simulation precision does not correlate with plausibility, as previously perceived in human evaluations. Therefore, optimizing LLM behavior based on human approval (e.g., using Reinforcement Learning from Human Feedback, RLHF) might not increase the precision of its explanations.

Implications and Future Directions

The implications of this research are multifold. Technically, it urges an improvement in LLM architectures and training paradigms that could more robustly interlink reasoning paths with output predictions. Practically, industries utilizing AI systems, particularly in explainable AI-required fields like healthcare and justice, must rethink current approbation-based methods (like RLHF) for models tasked with high-stakes decision-making.

For further research, extending these metrics to generation tasks predicates interest, given their broader applicability in nuanced domains where model outputs are not limited to binary outcomes. Additionally, the prospect of building mental models via iterative human-AI interactions could present avenues for achieving better counterfactual simulatability in systems deployed in dynamic, real-world environments.

In conclusion, the paper provides critical insights that challenge existing paradigms in AI explainability, arguing against an over-reliance on current methods of gaining human approval and urging for a deeper investigation into how explanations are formulated and received. This expands our horizon on model interpretability and usability, crucial for ethical and efficient AI integration into various domains.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Yanda Chen (13 papers)
  2. Ruiqi Zhong (27 papers)
  3. Narutatsu Ri (7 papers)
  4. Chen Zhao (249 papers)
  5. He He (71 papers)
  6. Jacob Steinhardt (88 papers)
  7. Zhou Yu (206 papers)
  8. Kathleen McKeown (85 papers)
Citations (39)