Decision Theoretic Foundations for Experiments Evaluating Human Decisions (2401.15106v5)
Abstract: How well people use information displays to make decisions is of primary interest in human-centered AI, model explainability, data visualization, and related areas. However, what constitutes a decision problem, and what is required for a study to establish that human decisions could be improved remain open to speculation. We propose a widely applicable definition of a decision problem synthesized from statistical decision theory and information economics as a standard for establishing when human decisions can be improved in HCI. We argue that to attribute loss in human performance to forms of bias, an experiment must provide participants with the information that a rational agent would need to identify the utility-maximizing decision. As a demonstration, we evaluate the extent to which recent evaluations of decision-making from the literature on AI-assisted decisions achieve these criteria. We find that only 10 (26\%) of 39 studies that claim to identify biased behavior present participants with sufficient information to characterize their behavior as deviating from good decision-making in at least one treatment condition. We motivate the value of studying well-defined decision problems by describing a characterization of performance losses they allow us to conceive. In contrast, the ambiguities of a poorly communicated decision problem preclude normative interpretation. We conclude with recommendations for practice.
- Applying quadratic scoring rule transparently in multiple choice settings: A note. Technical Report. Jena Economic Research Papers.
- Does the whole exceed its parts? the effect of ai explanations on complementary team performance. In Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems. 1–16.
- Dirk Bergemann and Stephen Morris. 2019. Information design: A unified perspective. Journal of Economic Literature 57, 1 (2019), 44–95.
- To trust or to think: cognitive forcing functions can reduce overreliance on AI in AI-assisted decision-making. Proceedings of the ACM on Human-Computer Interaction 5, CSCW1 (2021), 1–21.
- The role of explanations on trust and reliance in clinical decision support systems. In 2015 international conference on healthcare informatics. IEEE, 160–169.
- Colin Camerer. 1995. Individual decision making. Handbook of experimental economics (1995).
- Value-suppressing uncertainty palettes. In Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems. 1–11.
- Algorithm aversion: people erroneously avoid algorithms after seeing them err. Journal of Experimental Psychology: General 144, 1 (2015), 114.
- Evanthia Dimara and John Stasko. 2021. A critical reflection on visualization research: Where do decision making tasks hide? IEEE Transactions on Visualization and Computer Graphics 28, 1 (2021), 1128–1138.
- Uncertainty displays using quantile dotplots or cdfs improve transit decision-making. In Proceedings of the 2018 CHI conference on human factors in computing systems. 1–12.
- The impact of algorithmic risk assessments on human predictions and its analysis via crowdsourcing studies. Proceedings of the ACM on Human-Computer Interaction 5, CSCW2 (2021), 1–24.
- Tilmann Gneiting. 2011. Making and evaluating point forecasts. J. Amer. Statist. Assoc. 106, 494 (2011), 746–762.
- Daniel G Goldstein and David Rothschild. 2014. Lay understanding of probability distributions. Judgment and Decision making 9, 1 (2014), 1–14.
- Human evaluation of spoken vs. visual explanations for open-domain qa. arXiv preprint arXiv:2012.15075 (2020).
- Ben Green and Yiling Chen. 2019. Disparate interactions: An algorithm-in-the-loop analysis of fairness in risk assessments. In Proceedings of the conference on fairness, accountability, and transparency. 90–99.
- Visual reasoning strategies for effect size judgments and decisions. IEEE transactions on visualization and computer graphics 27, 2 (2020), 272–282.
- Bayesian-assisted inference from visualized data. IEEE Transactions on Visualization and Computer Graphics 27, 2 (2020), 989–999.
- A bayesian cognition approach to improve data visualization. In Proceedings of the 2019 chi conference on human factors in computing systems. 1–14.
- Tell me more? The effects of mental model soundness on personalizing an intelligent agent. In Proceedings of the sigchi conference on human factors in computing systems. 1–10.
- Towards a science of human-ai decision making: a survey of empirical studies. arXiv preprint arXiv:2112.11471 (2021).
- Towards a Science of Human-AI Decision Making: An Overview of Design Space in Empirical Human-Subject Studies. In Proceedings of the 2023 ACM Conference on Fairness, Accountability, and Transparency. 1369–1385.
- Nicolas S Lambert. 2011. Elicitation and evaluation of statistical forecasts. Preprint (2011).
- Optimization of scoring rules. In Proceedings of the 23rd ACM Conference on Economics and Computation. 988–989.
- Zhiyuan “Jerry” Lin, Jongbin Jung, Sharad Goel, and Jennifer Skeem. 2020. The limits of human predictions of recidivism. Science advances 6, 7 (2020), eaaz0652.
- Understanding the effect of out-of-distribution examples and interactive explanations on human-ai decision making. Proceedings of the ACM on Human-Computer Interaction 5, CSCW2 (2021), 1–45.
- Failures in contingent reasoning: The role of uncertainty. American Economic Review 109, 10 (2019), 3437–3474.
- C. Thi Nguyen. 2023. Value Collapse. Presented as the Annual Cardiff Lecture of the Royal Institute of Philosophy. https://royalinstitutephilosophy.org/event/value-collapse/
- The influence of different graphical displays on nonexpert decision making under uncertainty. Journal of Experimental Psychology: Applied 21, 1 (2015), 37.
- Understanding Uncertainty: How Lay Decision-makers Perceive and Interpret Uncertainty in Human-AI Decision Making. In Proceedings of the 28th International Conference on Intelligent User Interfaces. 379–396.
- Daniel Read. 2005. Monetary incentives, what are they good for? Journal of Economic Methodology 12, 2 (2005), 265–276.
- ” Why should i trust you?” Explaining the predictions of any classifier. In Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining. 1135–1144.
- Leonard J Savage. 1972. The foundations of statistics. Courier Corporation.
- Katie Steele and H Orri Stefánsson. 2015. Decision theory. (2015).
- John Von Neumann and Oskar Morgenstern. 2007. Theory of games and economic behavior (60th Anniversary Commemorative Edition). Princeton university press.
- Xinru Wang and Ming Yin. 2021. Are explanations helpful? a comparative study of the effects of explanations in ai-assisted decision-making. In 26th international conference on intelligent user interfaces. 318–328.
- The Rational Agent Benchmark for Data Visualization. IEEE Transactions on Visualization and Computer Graphics (forthcoming) (2023).
- How do visual explanations foster end users’ appropriate trust in machine learning?. In Proceedings of the 25th international conference on intelligent user interfaces. 189–201.
- Designing Shared Information Displays for Agents of Varying Strategic Sophistication. arXiv preprint arXiv:2310.10858 (2023).
- Sibyl: Understanding and addressing the usability challenges of machine learning in high-stakes decision making. IEEE Transactions on Visualization and Computer Graphics 28, 1 (2021), 1161–1171.