Papers
Topics
Authors
Recent
Search
2000 character limit reached

Controllable vs Uncontrollable Failure Attribution

Updated 11 April 2026
  • Controllable versus uncontrollable failure attribution distinguishes failures that can be mitigated through feasible interventions from those driven by immutable or exogenous factors.
  • The A2P and CAFA frameworks employ causal inference and counterfactual reasoning to pinpoint actionable inputs and optimize intervention policies in complex systems.
  • Empirical evaluations demonstrate enhanced accuracy and transparency in identifying and addressing failures, informing effective decision-making in fields like healthcare and multi-agent systems.

Controllable versus uncontrollable failure attribution refers to the systematic distinction between failures that could be prevented or mitigated via feasible interventions, and those that derive from factors fundamentally outside the scope of intervention. This distinction is central to debugging complex systems, interpreting predictive models in high-stakes domains such as medicine and public health, and operationalizing causal reasoning over multi-agent interactions. Methodologically, attributions of failure or risk can be partitioned along the axis of controllability, guiding both research on causal inference and the design of actionable, transparent decision-support tools.

1. Formal Definitions and Conceptual Foundations

In automated decision processes, machine learning, and multi-agent systems, the notion of controllability underpins the ability to identify where and how a system’s undesirable outcome could have been avoided via feasible changes. Two primary categorizations emerge:

  • Controllable failures: Adverse outcomes that would have been averted by a feasible, localized intervention (e.g., correcting an action, changing a policy, or modifying a controllable input).
  • Uncontrollable failures: Outcomes impervious to any such localized or practical intervention—arising from immutable features (age, genetic background), exogenous noise, or factors outside the modeled environment.

In feature attribution contexts, controllability demarcates the set of input features that an actor or policy-maker can in principle adjust (controllable features, FcF_c) from those which cannot be changed in practice (uncontrollable features, FuF_u) (Kovvuri et al., 2022). In causal sequential decision settings, controllability is assessed by whether single-step (or, more generally, minimal) interventions suffice to avert failure as verified through counterfactual reasoning (West et al., 12 Sep 2025).

2. Causal Frameworks and Mathematical Formalism

Multi-Agent Trajectory Models

In multi-agent systems, a trajectory is defined as

τ=[(s0,a0),(s1,a1),,(sT,aT)],\tau = [(s_0, a_0), (s_1, a_1), \ldots, (s_T, a_T)],

where sts_t is the global state at time tt and ata_t is the action taken. State transitions follow the structural equation

st+1=f(st,at,ϵt),s_{t+1} = f(s_t, a_t, \epsilon_t),

with ϵt\epsilon_t drawn from a distribution of unobserved exogenous factors (knowledge gaps, misperceptions, etc.).

The outcome function Z(τ){0,1}Z(\tau) \in \{0,1\} indicates success or failure, F[Z(τ)=1]F \equiv [Z(\tau)=1]. Interventions are formalized using Pearl’s do-operator: the trajectory after an intervention at time FuF_u0 is denoted FuF_u1, and system evolution proceeds as

FuF_u2

(West et al., 12 Sep 2025)

Feature Attribution Models

Given an input vector FuF_u3 with index set FuF_u4, controllable and uncontrollable features are defined respectively as those variables which can and cannot be altered by intervention. Traditional model-agnostic attributions (e.g., SHAP, LIME) do not distinguish between these sets, treating all inputs uniformly (Kovvuri et al., 2022).

3. Methodological Approaches for Attribution

Abduct–Act–Predict (A2P) Scaffolding

A2P is a causal inference framework for failure attribution in multi-agent conversational systems (West et al., 12 Sep 2025), operationalized as follows:

  1. Abduction: For a candidate failure step FuF_u5, infer the most plausible hidden cause FuF_u6 via posterior estimation:

FuF_u7

  1. Action: Identify a minimal corrective intervention FuF_u8 targeting FuF_u9:

τ=[(s0,a0),(s1,a1),,(sT,aT)],\tau = [(s_0, a_0), (s_1, a_1), \ldots, (s_T, a_T)],0

where τ=[(s0,a0),(s1,a1),,(sT,aT)],\tau = [(s_0, a_0), (s_1, a_1), \ldots, (s_T, a_T)],1 is an action-distance metric.

  1. Prediction: Simulate the outcome trajectory post-intervention, estimate τ=[(s0,a0),(s1,a1),,(sT,aT)],\tau = [(s_0, a_0), (s_1, a_1), \ldots, (s_T, a_T)],2. If τ=[(s0,a0),(s1,a1),,(sT,aT)],\tau = [(s_0, a_0), (s_1, a_1), \ldots, (s_T, a_T)],3, declare the failure at τ=[(s0,a0),(s1,a1),,(sT,aT)],\tau = [(s_0, a_0), (s_1, a_1), \ldots, (s_T, a_T)],4 controllable.

If no intervention at any τ=[(s0,a0),(s1,a1),,(sT,aT)],\tau = [(s_0, a_0), (s_1, a_1), \ldots, (s_T, a_T)],5 achieves τ=[(s0,a0),(s1,a1),,(sT,aT)],\tau = [(s_0, a_0), (s_1, a_1), \ldots, (s_T, a_T)],6, the failure is deemed uncontrollable.

Controllable fActor Feature Attribution (CAFA)

CAFA partitions input features into τ=[(s0,a0),(s1,a1),,(sT,aT)],\tau = [(s_0, a_0), (s_1, a_1), \ldots, (s_T, a_T)],7 and τ=[(s0,a0),(s1,a1),,(sT,aT)],\tau = [(s_0, a_0), (s_1, a_1), \ldots, (s_T, a_T)],8 and produces an attribution vector τ=[(s0,a0),(s1,a1),,(sT,aT)],\tau = [(s_0, a_0), (s_1, a_1), \ldots, (s_T, a_T)],9 with the key constraint sts_t0 for all sts_t1, reflecting only the influence of the controllable features (Kovvuri et al., 2022). CAFA proceeds through:

  • Selective perturbation: Only features in sts_t2 are perturbed around the input sts_t3 within a threshold sts_t4, with sts_t5 fixed for all sts_t6.
  • Surrogate modeling: Fit a high-capacity surrogate sts_t7 (e.g., random forest) on the perturbed dataset sts_t8.
  • Global-for-local SHAP explanation: Compute attributions using the SHAP method on sts_t9, guaranteeing uncontrollable features are assigned zero attribution due to invariance in tt0.

4. Empirical Evaluation and Case Studies

Multi-Agent Failure Attribution

On the WhoWhen benchmark for multi-agent dialogue failures, A2P achieves a step-level failure attribution accuracy of tt1 on the Algorithm-Generated subset, a tt2 improvement over the tt3 baseline. On the more complex Hand-Crafted subset, A2P attains tt4 (a tt5 gain over the tt6 baseline) (West et al., 12 Sep 2025). This demonstrates robust operationalization of controllable versus uncontrollable failures in sequential decision settings.

Feature Attribution in Medical and Public Health Applications

CAFA was evaluated on Simulacrum lung cancer and UCI breast-cancer datasets (Kovvuri et al., 2022). In both cases, CAFA suppresses attributions to uncontrollable features such as age or sex (Pearson correlation between CAFA and SHAP on controllable features tt7–tt8), and highlights controllable features (e.g., chemotherapy regimen, BMI). Applied to UK COVID-19 intervention data, CAFA attributed zero importance to uncontrollable indicators (cases, fatalities), and prioritized actionable intervention policies (e.g., public gathering bans, venue restrictions), consistent with established transmission risk factors.

5. Implications for Decision-Making and System Design

Partitioning failure attributions along controllability axes underpins actionable recommendations and root-cause analysis:

  • In medicine, the separation identifies which aspects of a risk profile or treatment outcome are potentially modifiable, shaping clinical intervention strategies.
  • In public health, the framework isolates policy levers (e.g., targeted restrictions) with maximal marginal effect, excluding immutable demographic or environmental variables from action planning.
  • In multi-agent and conversational AI, rigorous counterfactual analysis determines whether error remediation is feasible at the step/action level or if system/environmental limitations preclude prevention.

A clear delineation channels operational focus toward interventions that are expected to yield improvements, and avoids misplaced emphasis on intractable sources of failure.

6. Limitations and Open Questions

Several constraints and unresolved issues pertain to controllable vs. uncontrollable failure attribution:

  • Partitioning tt9 generally relies on a priori domain knowledge and may be nontrivial, especially as the set of “controllable” variables can evolve with technological advances (e.g., gene therapy potentially rendering genetic factors controllable) (Kovvuri et al., 2022).
  • The CAFA algorithm’s dependence on local perturbations and surrogate modeling may not extend straightforwardly to regression tasks or imbalance-heavy datasets. SHAP’s computational and statistical limitations (feature dependence, cost) are inherited by CAFA.
  • In A2P, the fidelity of predictive counterfactual simulations is bounded by the capabilities of the LLM, and the threshold ata_t0 for declaring controllability is heuristic (West et al., 12 Sep 2025).
  • Single-step or localized interventions are assumed; genuinely uncontrollable failures or those requiring coordinated multi-step interventions may be underrepresented.
  • Neither method quantifies the dominance of uncontrollable factors in the original model prediction—a possible direction for further risk decomposition (Kovvuri et al., 2022).
  • External, unmodeled exogenous factors may render some failures undiagnosable via language or feature-based approaches, motivating integration with external monitoring or simulation (West et al., 12 Sep 2025).

7. Comparative Summary Table

Aspect CAFA Approach (Kovvuri et al., 2022) A2P Scaffolding (West et al., 12 Sep 2025)
Domain Predictive modeling, XAI, feature attribution Multi-agent sequential decision, dialogue agents
Core notion of controllability Partition of input features (controllable/uncontrollable) Success probability under local action intervention
Method Selective feature perturbation, SHAP-based attribution Abduct–Act–Predict counterfactual routine
Output Attribution vector zeroed for uncontrollable features Step/action labeled as controllable/uncontrollable failure
Illustrative application Medicine, public health (cancer, COVID-19 policy) Root-cause error localization in dialogue systems

Both CAFA and A2P provide rigorous methodological frameworks for distinguishing controllable from uncontrollable attributions, enhancing transparency and verifiability in automated decision-support and failure analysis across scientific and engineering domains.

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Controllable Versus Uncontrollable Failure Attribution.