Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
91 tokens/sec
Gemini 2.5 Pro Premium
52 tokens/sec
GPT-5 Medium
24 tokens/sec
GPT-5 High Premium
28 tokens/sec
GPT-4o
85 tokens/sec
DeepSeek R1 via Azure Premium
87 tokens/sec
GPT OSS 120B via Groq Premium
478 tokens/sec
Kimi K2 via Groq Premium
221 tokens/sec
2000 character limit reached

Future of Work with AI Agents: Auditing Automation and Augmentation Potential across the U.S. Workforce (2506.06576v2)

Published 6 Jun 2025 in cs.CY, cs.AI, cs.CL, cs.HC, and cs.LG

Abstract: The rapid rise of compound AI systems (a.k.a., AI agents) is reshaping the labor market, raising concerns about job displacement, diminished human agency, and overreliance on automation. Yet, we lack a systematic understanding of the evolving landscape. In this paper, we address this gap by introducing a novel auditing framework to assess which occupational tasks workers want AI agents to automate or augment, and how those desires align with the current technological capabilities. Our framework features an audio-enhanced mini-interview to capture nuanced worker desires and introduces the Human Agency Scale (HAS) as a shared language to quantify the preferred level of human involvement. Using this framework, we construct the WORKBank database, building on the U.S. Department of Labor's O*NET database, to capture preferences from 1,500 domain workers and capability assessments from AI experts across over 844 tasks spanning 104 occupations. Jointly considering the desire and technological capability divides tasks in WORKBank into four zones: Automation "Green Light" Zone, Automation "Red Light" Zone, R&D Opportunity Zone, Low Priority Zone. This highlights critical mismatches and opportunities for AI agent development. Moving beyond a simple automate-or-not dichotomy, our results reveal diverse HAS profiles across occupations, reflecting heterogeneous expectations for human involvement. Moreover, our study offers early signals of how AI agent integration may reshape the core human competencies, shifting from information-focused skills to interpersonal ones. These findings underscore the importance of aligning AI agent development with human desires and preparing workers for evolving workplace dynamics.

Summary

  • The paper presents WORKBank, a survey-based auditing framework that evaluates AI agent automation and augmentation across 844 detailed occupational tasks.
  • It reveals significant mismatches between worker desires and expert capability assessments, emphasizing the need for targeted R&D in AI development.
  • The study highlights a shift in core skills from information processing to interpersonal competence, underscoring the importance of workforce retraining.

Auditing Automation and Augmentation Potential of AI Agents

This paper (2506.06576) introduces a survey-based auditing framework, WORKBank, to evaluate the potential of AI agents in automating or augmenting tasks across the U.S. workforce. The framework captures worker desires and expert assessments, incorporating the Human Agency Scale (HAS) to quantify human involvement. The paper's core contribution lies in its comprehensive analysis of the desire-capability landscape, revealing mismatches and opportunities for AI agent development, while also highlighting the potential shift in core human skills due to AI integration.

Framework and Methodology

The auditing framework (Figure 1) centers on complex, multi-step tasks sourced from the O*NET database and emphasizes the spectrum of automation and augmentation. The HAS, a five-level scale ranging from full automation (H1) to essential human involvement (H5) (Figure 2), provides a human-centered lens for assessing task properties and guiding agent development. The framework collects worker ratings on automation desire and desired HAS levels, using an audio-enhanced survey system. AI experts then assess automation capability and feasible HAS levels. The WORKBank database, built upon this framework, consists of responses from 1,500 workers and 52 AI experts, covering 844 occupational tasks. Figure 1

Figure 1: Overview of the auditing framework and key insights.

Figure 2

Figure 2: Levels of Human Agency Scale (HAS).

Key Findings and Analysis

The paper reveals several key findings:

  • Workers desire automation for low-value and repetitive tasks. For 46.1% of tasks, workers express positive attitudes toward AI agent automation, motivated primarily by freeing up time for high-value work (Figure 3).
  • The desire-capability landscape reveals critical mismatches. Tasks are divided into four zones: Automation "Green Light," "Red Light," R{content}D Opportunity, and Low Priority. Investments are disproportionately concentrated in the Low Priority and Automation "Red Light" Zones (Figure 4).
  • The HAS provides a shared language to audit AI use at work. Workers generally prefer higher levels of human agency than experts deem necessary (Figure 5).
  • Key human skills may be shifting from information processing to interpersonal competence (Figure 6).

Implications and Future Directions

These findings have significant implications for AI agent research, development, and deployment strategies. The identified mismatches between worker desires and technological capabilities highlight the need for more targeted R{content}D efforts. The HAS framework offers a valuable tool for understanding and promoting human-agent collaboration, while the observed shift in core human skills underscores the importance of workforce development and retraining initiatives. Future research should focus on longitudinal tracking of task-level changes and further exploration of worker reskilling strategies.

Critical Assessment

The paper presents a comprehensive and well-executed paper of the potential impact of AI agents on the U.S. workforce. The auditing framework and WORKBank database represent valuable resources for researchers and practitioners in the field. The analysis of the desire-capability landscape and the HAS framework provide novel insights into the complex interplay between automation and augmentation.

One potential limitation is the reliance on existing occupational tasks defined by the O*NET database, which may not fully capture the emergence of new tasks as AI agents become more integrated into the workplace. Additionally, while the paper includes a diverse sample of workers, it is possible that participants' responses may be influenced by their limited awareness of AI capabilities and constraints, as well as concerns about job security.

Conclusion

This paper makes a significant contribution to the understanding of the future of work with AI agents. By integrating worker desires and expert assessments, the paper provides actionable insights for aligning AI development with human needs and preparing workers for evolving workplace dynamics. The WORKBank database and auditing framework offer a valuable foundation for future research and policy initiatives in this rapidly evolving field.

Youtube Logo Streamline Icon: https://streamlinehq.com