Human Attribute Taxonomy for Hybrid Systems
- Human-Attribute Taxonomy is a framework that defines 12 distinct axes differentiating human and ML decision-making processes.
- It categorizes differences into task, input, internal processing, and output stages, enabling a structured mathematical aggregation of decisions.
- The taxonomy informs hybrid system design by leveraging agent complementarity to achieve joint performance that outperforms standalone approaches.
A human-attribute taxonomy systematically characterizes the distinct axes along which human and ML agents differ in decision-making systems, providing a conceptual foundation for understanding and engineering human–ML complementarity in @@@@1@@@@ decision contexts. This taxonomy enumerates 12 attributes, grouped along four computational-level stages—task definition, input, internal processing, and output. Each attribute formalizes a specific source of discrepancy between human and ML agents, together forming the basis for mathematical analysis of joint performance and the conditions under which hybrid systems can surpass either agent alone (Rastogi et al., 2022).
1. Taxonomy of Human vs. ML Decision-Making Attributes
The taxonomy comprises 12 attributes, each precisely defined, rationalized, and notated, indicating how humans and ML agents diverge in operational mechanisms and decision heuristics.
| Attribute | Notation | Example Context |
|---|---|---|
| Objective | Human loan officer vs. ML risk minimization | |
| Misaligned Construct of Interest | Medical necessity vs. historic spending | |
| Access to Different Information | Judge’s perception vs. ML coded data | |
| Nature of Past Experiences | Physician’s clinical history vs. ML dataset | |
| Models of the World | Human rules-of-thumb vs. NN feature maps | |
| Input Processing and Perception | Probability distortion, causal/statistical bias | |
| Choosing Among Models (Optimization) | Human heuristics vs. ML global optimization | |
| Available Actions | Human flexibility vs. ML prescription | |
| Explaining the Decision | — | Social narrative vs. model saliency map |
| Uncertainty Communication | (ML), (human) | Probabilistic output vs. discrete confidence |
| Output Consistency | Human noise vs. ML determinism | |
| Time Efficiency | ML high throughput vs. human limits |
Task Definition differences comprise objective misalignment and divergence between constructs of interest, reflecting that ML agents usually optimize explicit, mathematically tractable objectives, whereas humans institute broader, context-sensitive evaluative functions ( vs. ). ML models typically learn to optimize a proxy , while human agents reason about a latent, true construct .
Input Differences pertain to disparities in observed feature spaces () and the nature or breadth of prior experience (), revealing that lived experience and unstructured observation contribute to divergences in judgment.
Internal Processing Divergences are articulated through policy classes (), perception mappings (), and the adopted optimization strategies (human heuristics OPT versus ML global search OPT). These highlight differences in causal versus statistical reasoning, bounded rationality, and algorithmic constraints.
Output Heterogeneity encompasses not only the actions space () but also differences in explanation modalities, uncertainty communication, temporal consistency, and efficiency metrics.
2. Mathematical Aggregation Framework Formalizing Complementarity
A convex combination framework formalizes hybrid human–ML decision-making, enabling precise analysis of when and how agent complementarity emerges. Each agent supplies a local policy, evaluated on a global feature space , with possible decisions in action space .
Given decision instances , the aggregation constructs a weighted policy:
- and are per-instance mixing weights.
Optimal aggregation chooses the weight vector to maximize the target score:
where is the evaluation function.
A system exhibits complementarity if:
This quantifies joint-system value beyond the best stand-alone agent.
Collaboration mechanisms are further distinguished:
- Across-instance complementarity (AC):
where . AC measures routing/deferral frequency.
- Within-instance complementarity (WC):
WC is maximal for full blending () across all , and zero for pure routing.
3. Mechanisms Underlying Human–ML Complementarity
Complementarity arises specifically when humans and ML agents differ along one or more taxonomy attributes. The key mechanisms identified include:
- Input Asymmetry (): Each agent may be strong on distinct features, favoring local averaging (high ) when features are complementary, or pure deferral () when a clear strength dominates per instance.
- Objective Misalignment (): Aggregation should route occasions to the agent best aligned with the target evaluation function , especially when each agent’s training involved a different cost function or label proxy.
- Model Class Discrepancy (): Human mental models may generalize out-of-distribution where ML models excel at exploiting high-volume statistical structure; aggregation can blend their generalization domains.
- Inconsistency/Noise: Human decisions are susceptible to intra-individual variability (“noise”), whereas ML agents tend toward deterministic output. The aggregation weights can be adapted dynamically to prioritize ML when human inconsistency is detected.
- Time Constraints: Trivial or routine decisions can be assigned to ML for efficiency, whereas exceptions or high-uncertainty cases are deferred to human agents.
Each mechanism can be explicitly mapped to the manner in which aggregation weights are learned and deployed over instances, with performance gains explained by the specific diversity of strengths engaged.
4. Empirical and Theoretical Conditions for Gaining Complementarity
Synthetic simulations in the referenced work and general theoretical analysis confirm that the prerequisite for sustainable complementarity is persistent, exploitable difference along taxonomy axes. Hybrid systems that simply mirror one another or are identical in policy class, training data, or objectives, can at best match, but not exceed, standalone performance.
Critical conditions include:
- Sufficient input diversity (distinct information or lived experience).
- Persistent optimization and model class differences (causal vs. statistical, heuristic vs. global optimality).
- Non-negligible discrepancy in action space, explanation, and uncertainty signaling.
- Task/objective misalignment between agents.
These conditions enable aggregation policies to leverage both deferral and blending, aligning the chosen agent or combination to the local structure and risk of each decision.
5. Implications for Hybrid System Design and Future Research
The taxonomy provides a principled language and analytic foundation to reason about the sources of human–ML complementarity, supporting both empirical inquiry and system design. Its explicit mapping between attribute differences and aggregation mechanisms allows designers to predict—not merely hope for—complementarity by probing which differences are likely to be relevant for the target domain.
A plausible implication is that attempts to maximize complementarity should focus on cultivating and preserving such differences rather than minimizing human–ML discrepancies. Redundant overlap in agent knowledge or identical optimization strategies can reduce hybrid gains; by contrast, heterogeneity along the taxonomy’s axes offers the raw material for outperforming stand-alone systems.
6. Synthesis: Role of the Human-Attribute Taxonomy in Complementarity
In summary, the taxonomy of human versus ML attributes provides a rigorous scaffold for analyzing, engineering, and optimizing hybrid decision-making systems. Each attribute—whether it pertains to objectives, information, model class, action space, or output modalities—serves as both a diagnostic lens and a design lever. The convex-combination aggregation framework mathematically formalizes how these diverse strengths can be strategically combined to yield true complementarity, as determined by strict improvement over either agent (Eq. 3). Together, these scientific tools advance the understanding and deployment of hybrid human–ML systems across consequential domains (Rastogi et al., 2022).