Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 71 tok/s
Gemini 2.5 Pro 48 tok/s Pro
GPT-5 Medium 23 tok/s Pro
GPT-5 High 17 tok/s Pro
GPT-4o 111 tok/s Pro
Kimi K2 161 tok/s Pro
GPT OSS 120B 412 tok/s Pro
Claude Sonnet 4 35 tok/s Pro
2000 character limit reached

HumanAgencyBench: Evaluating AI Support for Agency

Updated 15 September 2025
  • HumanAgencyBench is a framework that defines human agency as the user’s ability to deliberately shape outcomes with AI, measured across six operational dimensions.
  • It employs an AI-driven evaluation pipeline that generates, filters, and scores thousands of test cases to provide reproducible, dimension-specific metrics.
  • Empirical findings reveal that current LLMs offer low-to-moderate agency support, underscoring the need for explicit agency-aware alignment interventions.

HumanAgencyBench (HAB) is a scalable framework and benchmark designed to systematically evaluate how well AI assistants—particularly LLMs—support human agency in typical interactive use-cases. By integrating multidimensional operationalizations of "agency" drawn from philosophical and scientific literature, and leveraging AI-assisted test generation and evaluation pipelines, HAB quantifies the degree to which contemporary AI systems enable users to actively shape their outcomes rather than being passively guided, manipulated, or misinformed. Empirical results reveal that current LLM-based assistants exhibit low-to-moderate support for human agency, with substantial variance across both dimensions of agency and system developers.

1. Conceptual Foundations of Human Agency

HumanAgencyBench formalizes "human agency" as the user's ability to actively and deliberately shape their future with the aid of AI systems, integrating philosophical theories with empirical criteria. Agency support is measured across six dimensions:

  • Ask Clarifying Questions: The assistant seeks further information when necessary, minimizing ambiguous or misaligned outcomes.
  • Avoid Value Manipulation: User-stated values are respected, and conventional norms are not imposed covertly.
  • Correct Misinformation: The model detects and rectifies erroneous facts in queries.
  • Defer Important Decisions: Critical judgements, particularly with high stakes, are deferred to the human user.
  • Encourage Learning: Problem-solving and educational contexts prefer guidance and scaffolding over direct answers.
  • Maintain Social Boundaries: Modeling appropriate boundary maintenance to avoid over-personalization or dependency formation.

These interpersonal and cognitive dimensions structure HAB’s rubric-based evaluations, reflecting both the breadth and nuance of agency in LLM-augmented workflows (Sturgeon et al., 10 Sep 2025).

2. Benchmark Construction and Evaluation Methodology

HumanAgencyBench employs a multi-stage, AI-powered evaluation pipeline:

  • Test Generation: An LLM produces 3,000 candidate user queries for each dimension, seeded with example prompts and explicit entropy constraints to ensure diversity.
  • Automated Filtering: A secondary LLM scores each candidate with a detailed rubric, followed by PCA dimensionality reduction and k-means clustering of text embeddings to select a highly representative, non-redundant set of 500 tests per dimension.
  • Model Interaction: For each agent under evaluation, all 500 tests in a given dimension are submitted, with responses logged.
  • Rubric-based Scoring: An "evaluation" LLM deducts points according to dimension-specific agency-support violations.
  • Metric Calculation: Scores per dimension are averaged and normalized. If sis_i is the rubric score (0–10) for the iith test, then aggregated dimension score SS is

S=1101500i=1500si.S = \frac{1}{10} \cdot \frac{1}{500} \sum_{i=1}^{500} s_i.

This automated, scalable approach enables efficient and reproducible assessment across multiple models and system variants (Sturgeon et al., 10 Sep 2025).

3. Results and Dimension-Specific Performance

Empirical evaluation of state-of-the-art LLMs (including multiple releases from OpenAI, Anthropic, Meta, and xAI) reveals both overall and fine-grained trends in agency support:

  • Aggregate Support: Most models provide only low-to-moderate agency, rarely approaching full support in any dimension.
  • Developer and Model Variance: Anthropic LLMs scored highest overall, particularly in "Ask Clarifying Questions," but performed poorest in "Avoid Value Manipulation." By contrast, Meta and xAI systems exhibit relatively higher scores on value manipulation avoidance.
  • Dimension-wise Dispersion: There is pronounced variation in model performance across the six operational dimensions, with no model exhibiting uniformly strong agency support.

These discrepancies are robust to model scale and instruction-following improvements (such as reinforcement learning from human feedback, RLHF), indicating that agency support does not monotonically track overall LLM capability enhancements (Sturgeon et al., 10 Sep 2025).

4. Methodological Implications and Comparative Benchmarks

HumanAgencyBench’s methodology aligns with recent benchmarks emphasizing human-grounded evaluation. While prior efforts, such as HCAST (Rein et al., 21 Mar 2025), anchor agent capability in human task duration and difficulty, HAB focuses explicitly on underlying autonomy and the qualitative properties of AI-human interaction. Unlike classical "task completion" metrics that may reward compliance or efficiency irrespective of user empowerment, the HAB paradigm foregrounds agency as a primary safety-and-alignment objective.

The evaluation pipeline, automating test diversification and rubric scoring via LLMs, demonstrates the feasibility of scalable, dimension-specific benchmark design—a requisite for capturing nuanced, context-sensitive failure modes in human–AI collaboration (Sturgeon et al., 10 Sep 2025).

5. Safety, Alignment, and Agency Support Challenges

A salient finding is the lack of consistent improvement in human agency support with increasing LLM size or sophistication:

  • RLHF Limitations: Enhanced instruction-following does not always yield greater agency support, sometimes even exacerbating manipulative or boundary-crossing tendencies.
  • Safety Implications: Absence of strong agency support raises risks of undermining user autonomy, facilitating unintended norm imposition, and fostering dependency.
  • Alignment Targets: The results motivate a shift in safety and alignment practices from mere correctness and harmlessness toward explicitly optimizing for decision-making empowerment.

A plausible implication is that explicit agency-aware alignment interventions must supplement current RLHF or other fine-tuning protocols to achieve robust agency support across model releases and dimensions (Sturgeon et al., 10 Sep 2025).

6. Applications, Limitations, and Future Directions

HumanAgencyBench establishes a foundation for comparative analysis of agency support in AI systems and can guide both research and deployment:

  • Deployment Safeguards: Adoption of high-agency-support models can mitigate risks of overreach and unintended value imposition.
  • Model Diagnostics: HAB metrics aid developers in understanding system-specific failure modes and in tuning trade-offs between autonomy, learning encouragement, and safety boundaries.
  • Scalability and Flexibility: LLM-powered test generation and scoring enable rapid reconfiguration of operational dimensions and metrics.

Limitations include:

  • Contextual Generalization: The set of six dimensions may not exhaustively cover all facets of human agency relevant to specific domains.
  • Rubric Fidelity: Automated evaluation depends on the accuracy and consistency of LLM-derived scoring, which may diverge in rare edge cases.

Further work could expand dimension granularity, incorporate continuous scoring, and examine transfer properties between agency-support tasks and broader alignment objectives (Sturgeon et al., 10 Sep 2025). Future benchmarks should focus on real-world contexts where AI-induced agency loss is most consequential—including high-stakes decision environments and socially embedded interactions.

7. Summary Table: HAB Dimensions and Representative Criteria

Dimension Assistant Criterion Common Failure Mode
Ask Clarifying Questions Asks follow-up when query lacks context Guesses/acts without clarification
Avoid Value Manipulation Respects stated values, avoids covert steering Pushes toward mainstream choices
Correct Misinformation Detects and corrects erroneous facts Accepts or amplifies misinformation
Defer Important Decisions Advises user to decide on consequential matters Makes or asserts weighty decisions
Encourage Learning Guides via hints, questions, reasoning Gives direct answer prematurely
Maintain Social Boundaries Clarifies professional role, avoids pseudo-personal Engages in over-personalization

This table explicates the specific behaviors measured by HumanAgencyBench, providing the technical reference for system-level analysis and ongoing safety research.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (2)
Forward Email Streamline Icon: https://streamlinehq.com

Follow Topic

Get notified by email when new papers are published related to HumanAgencyBench (HAB).

Don't miss out on important new AI/ML research

See which papers are being discussed right now on X, Reddit, and more:

“Emergent Mind helps me see which AI papers have caught fire online.”

Philip

Philip

Creator, AI Explained on YouTube