Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Human Learning about AI Performance (2406.05408v1)

Published 8 Jun 2024 in econ.GN and q-fin.EC

Abstract: How do humans assess the performance of AI across different tasks? AI has been noted for its surprising ability to accomplish very complex tasks while failing seemingly trivial ones. We show that humans engage in ``performance anthropomorphism'' when assessing AI capabilities: they project onto AI the ability model that they use to assess humans. In this model, observing an agent fail an easy task is highly diagnostic of a low ability, making them unlikely to succeed at any harder task. Conversely, a success on a hard task makes successes on any easier task likely. We experimentally show that humans project this model onto AI. Both prior beliefs and belief updating about AI performance on standardized math questions appear consistent with the human ability model. This contrasts with actual AI performance, which is uncorrelated with human difficulty in our context, and makes such beliefs misspecified. Embedding our framework into an adoption model, we show that patterns of under- and over-adoption can be sustained in an equilibrium with anthropomorphic beliefs.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (2)
  1. Bnaya Dreyfuss (1 paper)
  2. Raphael Raux (1 paper)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com