Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Trust Your Gut: Comparing Human and Machine Inference from Noisy Visualizations (2407.16871v1)

Published 23 Jul 2024 in cs.HC and cs.LG

Abstract: People commonly utilize visualizations not only to examine a given dataset, but also to draw generalizable conclusions about the underlying models or phenomena. Prior research has compared human visual inference to that of an optimal Bayesian agent, with deviations from rational analysis viewed as problematic. However, human reliance on non-normative heuristics may prove advantageous in certain circumstances. We investigate scenarios where human intuition might surpass idealized statistical rationality. In two experiments, we examine individuals' accuracy in characterizing the parameters of known data-generating models from bivariate visualizations. Our findings indicate that, although participants generally exhibited lower accuracy compared to statistical models, they frequently outperformed Bayesian agents, particularly when faced with extreme samples. Participants appeared to rely on their internal models to filter out noisy visualizations, thus improving their resilience against spurious data. However, participants displayed overconfidence and struggled with uncertainty estimation. They also exhibited higher variance than statistical machines. Our findings suggest that analyst gut reactions to visualizations may provide an advantage, even when departing from rationality. These results carry implications for designing visual analytics tools, offering new perspectives on how to integrate statistical models and analyst intuition for improved inference and decision-making. The data and materials for this paper are available at https://osf.io/qmfv6

Human and Machine Inference from Noisy Visualizations

The paper "Trust Your Gut: Comparing Human and Machine Inference from Noisy Visualizations" addresses the complex task of making inferences from visual data under conditions of uncertainty, comparing human judgment with Bayesian statistical models. The authors, Koonchanok, Papka, and Reda, scrutinize whether human intuition can outperform Bayesian inference in particular situations, acknowledging that humans, while generally less rational, may provide superior assessments under certain conditions.

Key Findings and Methodology

The research comprises two primary experiments designed to evaluate the inference accuracy of humans compared to Bayesian agents when interpreting visualizations depicting bivariate relationships. These experiments manipulate variables such as visualization type, sample size, sample extremeness, and levels of uncertainty to draw robust conclusions.

  1. Experiment I: This experiment investigates how visualization type, sample size, and social consensus impact human inference accuracy compared to Bayesian agents. Notably, human analysts were found to outperform Bayesian agents when samples were extreme, implying that in cases of spurious data, intuitive human judgment may resist misleading information better than statistical models.
  2. Experiment II: This experiment focuses on uncertainty levels in data-generation processes and their influence on inference accuracy. While participants did not close the accuracy gap with statistical machines, increased social consensus and lower uncertainty consistently improved human performance independently.

Implications of Human Versus Machine Inference

The findings emphasize that humans exhibit a particular robustness to extreme data samples which Bayesian agents frequently misinterpret. This insight into human intuition's capability to discount outlier information points to the potential of harnessing such intuitive processes within human-machine collaboration frameworks. The research suggests that designing visualization systems to encourage leveraging human intuition, particularly in exploratory analysis, may lead to more effective outcomes by preventing false discoveries.

These insights encourage a reconsideration of how humans and algorithms can coalesce for optimal data analysis outcomes. Instead of striving for purely normative, Bayesian approaches, visual analytics systems might benefit from a symbiotic integration of human heuristics and machine precision, wherein humans handle smaller, more ambiguous datasets, and machines are used for large-volume data interpretation.

Challenges and Future Directions

Despite these findings, the variability in human responses and their overconfidence in uncertainty assessment remain challenges. Future studies could explore refining elicitation interfaces and examining more extensive datasets and complex, expertise-dependent topics to better understand and harness this variability.

The work prompts a deeper inquiry into collaborative inference models that blend human intuition and machine algorithms, potentially leading to the development of adaptive visualization systems. Such systems could enhance the decision-making process in data analytics, particularly in environments with significant uncertainties, furthering the field of human-computer interaction in data science.

In summary, the research presented in this paper reveals that the non-rational elements of human inference can offer valuable advantages when complemented with Bayesian statistical models, advocating for a more nuanced approach to human-AI collaboration in data analysis.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (71)
  1. How can ai automate end-to-end data science? arXiv preprint arXiv:1910.14436, 2019. doi: 10 . 48550/arXiv . 1910 . 14436
  2. Can intuition improve deception detection performance? Journal of Experimental Social Psychology, 45(4):1052–1055, 2009. doi: 10 . 1016/j . jesp . 2009 . 05 . 017
  3. Methods for discovering cognitive biases in a visual analytics environment. Cognitive Biases in Visualizations, pp. 61–73, 2018. doi: 10 . 1007/978-3-319-95831-6_5
  4. Map lineups: Effects of spatial structure on graphical inference. IEEE Transactions on Visualization and Computer Graphics, 23(1):391–400, 2016. doi: 10 . 1109/TVCG . 2016 . 2598862
  5. Role of human-ai interaction in selective prediction. In Proceedings of the AAAI Conference on Artificial Intelligence, vol. 36, pp. 5286–5294, 2022. doi: 10 . 1609/aaai . v36i5 . 20465
  6. Statistical inference for exploratory data analysis and model diagnostics. Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences, 367(1906):4361–4383, 2009. doi: 10 . 1098/rsta . 2009 . 0120
  7. P.-C. Bürkner. brms: An r package for bayesian multilevel models using stan. Journal of statistical software, 80:1–28, 2017. doi: 10 . 18637/jss . v080 . i01
  8. C. F. Camerer and H. Kunreuther. Decision processes for low probability events: Policy implications. Journal of Policy Analysis and Management, 8(4):565–592, 1989. doi: 10 . 2307/3325045
  9. The anchoring effect in decision-making with visual analytics. In 2017 IEEE Conference on Visual Analytics Science and Technology (VAST), pp. 116–126. IEEE, 2017. doi: 10 . 1109/VAST . 2017 . 8585665
  10. Concept-driven visual analytics: an exploratory study of model-and hypothesis-based reasoning with visualizations. In Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems, pp. 1–14, 2019. doi: 10 . 1145/3290605 . 3300298
  11. Visual (dis) confirmation: Validating models and hypotheses with visualizations. In 2019 23rd International Conference in Information Visualization–Part II, pp. 116–121. IEEE, 2019. doi: 10 . 1109/IV-2 . 2019 . 00032
  12. Bayesian methods in extreme value modelling: a review and new developments. International Statistical Review/Revue Internationale de Statistique, pp. 119–136, 1996. doi: 10 . 2307/1403426
  13. A bayesian analysis of extreme rainfall data. Journal of the Royal Statistical Society Series C: Applied Statistics, 45(4):463–478, 1996. doi: 10 . 2307/2986068
  14. V. Crupi and F. Calzavarini. Critique of pure bayesian cognitive science: A view from the philosophy of science. European Journal for Philosophy of Science, 13(3):28, 2023. doi: 10 . 1007/s13194-023-00533-w
  15. W. Cui. Visual analytics: A comprehensive overview. IEEE Access, 7:81555–81573, 2019. doi: 10 . 1109/ACCESS . 2019 . 2923736
  16. When should i trust my gut? linking domain expertise to intuitive decision-making effectiveness. Organizational behavior and human decision processes, 119(2):187–194, 2012. doi: 10 . 1016/j . obhdp . 2012 . 07 . 009
  17. A task-based taxonomy of cognitive biases for information visualization. IEEE Transactions on Visualization and Computer Graphics, 26(2):1413–1432, 2018. doi: 10 . 1109/TVCG . 2018 . 2872577
  18. Trust in automl: exploring information needs for establishing trust in automated machine learning systems. In Proceedings of the 25th international conference on intelligent user interfaces, pp. 297–307, 2020. doi: 10 . 1145/3377325 . 3377501
  19. The human is the loop: new directions for visual analytics. Journal of Intelligent Information Systems, 43:411–435, 2014. doi: 10 . 1007/s10844-014-0304-9
  20. Detecting and avoiding likely false-positive findings–a practical guide. Biological Reviews, 92(4):1941–1968, 2017. doi: 10 . 1111/brv . 12315
  21. Cognitive challenges in human–artificial intelligence collaboration: Investigating the path toward productive delegation. Information Systems Research, 33(2):678–696, 2022. doi: 10 . 1287/isre . 2021 . 1079
  22. Using icon arrays to communicate medical risks: overcoming low numeracy. Health psychology, 28(2):210, 2009. doi: 10 . 1037/a0014474
  23. Do icon arrays help reduce denominator neglect? Medical Decision Making, 30(6):672–684, 2010. doi: 10 . 1177/0272989X10369000
  24. Bayesian workflow. arXiv preprint arXiv:2011.01808, 2020. doi: 10 . 48550/arXiv . 2011 . 01808
  25. G. Gigerenzer. Why heuristics work. Perspectives on psychological science, 3(1):20–29, 2008. doi: 10 . 1111/j . 1745-6916 . 2008 . 00058 . x
  26. G. Gigerenzer and H. Brighton. Homo heuristicus: Why biased minds make better inferences. Topics in Cognitive Science, 1(1):107–143, 2009. doi: 10 . 1111/j . 1756-8765 . 2008 . 01006 . x
  27. G. Gigerenzer and W. Gaissmaier. Heuristic decision making. Annual review of psychology, 62:451–482, 2011. doi: 10 . 1146/annurev-psych-120709-145346
  28. G. Gigerenzer and P. M. Todd. Fast and frugal heuristics: The adaptive toolbox. In Simple heuristics that make us smart, pp. 3–34. Oxford University Press, 1999.
  29. Heuristics and biases: The psychology of intuitive judgment. Cambridge university press, 2002. doi: 10 . 1017/CBO9780511808098
  30. Ranking visualizations of correlation using weber’s law. IEEE Transactions on Visualization and Computer Graphics, 20(12):1943–1952, 2014. doi: 10 . 1109/TVCG . 2014 . 2346979
  31. Adaptive rationality: An evolutionary perspective on cognitive bias. Social Cognition, 27(5):733–763, 2009. doi: 10 . 1521/soco . 2009 . 27 . 5 . 733
  32. Human-ai collaboration: The effect of ai delegation on human task performance and task satisfaction. In Proceedings of the 28th International Conference on Intelligent User Interfaces, pp. 453–463, 2023. doi: 10 . 1145/3581641 . 3584052
  33. Pushing the (visual) narrative: the effects of prior knowledge elicitation in provocative topics. In Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems, pp. 1–14, 2020. doi: 10 . 1145/3313831 . 3376887
  34. L. Huang. The role of investor gut feel in managing complexity and extreme risk. Academy of Management Journal, 61(5):1821–1847, 2018. doi: 10 . 5465/amj . 2016 . 1009
  35. Judgment under uncertainty: Heuristics and biases. Cambridge university press, 1982.
  36. Evm: Incorporating model checking into exploratory visual analysis. IEEE Transactions on Visualization and Computer Graphics, 2023. doi: 10 . 1109/TVCG . 2023 . 3326516
  37. A bayesian cognition approach for belief updating of correlation judgement through uncertainty visualizations. IEEE Transactions on Visualization and Computer Graphics, 27(2):978–988, 2020. doi: 10 . 1109/TVCG . 2020 . 3029412
  38. Automl to date and beyond: Challenges and opportunities. ACM Computing Surveys (CSUR), 54(8):1–36, 2021. doi: 10 . 1145/3470918
  39. Bayesian-assisted inference from visualized data. IEEE Transactions on Visualization and Computer Graphics, 27(2):989–999, 2020. doi: 10 . 1109/TVCG . 2020 . 3028984
  40. A bayesian cognition approach to improve data visualization. In Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems, pp. 1–14, 2019. doi: 10 . 1145/3290605 . 3300912
  41. J. J. Koehler and L. Macchi. Thinking about low-probability events: An exemplar-cuing theory. Psychological Science, 15(8):540–546, 2004. doi: 10 . 1111/j . 0956-7976 . 2004 . 00716 . x
  42. Data prophecy: Exploring the effects of belief elicitation in visual analytics. In Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems, pp. 1–12, 2021. doi: 10 . 1145/3411764 . 3445798
  43. Visual belief elicitation reduces the incidence of false discovery. In Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems, pp. 1–17, 2023. doi: 10 . 1145/3544548 . 3580808
  44. Human-ai collaboration via conditional delegation: A case study of content moderation. In Proceedings of the 2022 CHI Conference on Human Factors in Computing Systems, pp. 1–18, 2022. doi: 10 . 1145/3491102 . 3501999
  45. Scientific productivity: An exploratory study of metrics and incentives. PloS one, 13(4):e0195321, 2018. doi: 10 . 1371/journal . pone . 0195321
  46. Ecological rationality: Fast-and-frugal heuristics for managerial decision making under uncertainty. Academy of Management Journal, 62(6):1735–1759, 2019. doi: 10 . 5465/amj . 2018 . 0172
  47. Improving inferences in population studies of rare species that are detected imperfectly. Ecology, 86(5):1101–1113, 2005. doi: 10 . 1890/04-1060
  48. Vibe: A design space for visual belief elicitation in data journalism. In Computer Graphics Forum, vol. 41, pp. 477–488. Wiley Online Library, 2022. doi: 10 . 1111/cgf . 14556
  49. Validation of visual statistical inference, applied to linear models. Journal of the American Statistical Association, 108(503):942–956, 2013. doi: 10 . 1080/01621459 . 2013 . 808157
  50. When do data visualizations persuade? the impact of prior attitudes on learning about correlations from scatterplot visualizations. In Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems, pp. 1–16, 2023. doi: 10 . 1145/3544548 . 3581330
  51. Assessing the effect of visualizations on bayesian reasoning through crowdsourcing. IEEE Transactions on Visualization and Computer Graphics, 18(12):2536–2545, 2012. doi: 10 . 1109/TVCG . 2012 . 199
  52. Teaching humans when to defer to a classifier via exemplars. In Proceedings of the AAAI Conference on Artificial Intelligence, vol. 36, pp. 5323–5331, 2022. doi: 10 . 48550/arXiv . 2111 . 11297
  53. Scientific utopia: Ii. restructuring incentives and practices to promote truth over publishability. Perspectives on Psychological Science, 7(6):615–631, 2012. doi: 10 . 1177/1745691612459058
  54. Improving bayesian reasoning: The effects of phrasing, visualization, and spatial ability. IEEE transactions on visualization and computer graphics, 22(1):529–538, 2015. doi: 10 . 1109/TVCG . 2015 . 2467758
  55. Ai knowledge: Improving ai delegation through human enablement. In Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems, pp. 1–17, 2023. doi: 10 . 1145/3544548 . 3580794
  56. Modeling and evaluating user behavior in exploratory visual analysis. Information Visualization, 15(4):325–339, 2016. doi: 10 . 1177/1473871616638546
  57. The impact of elicitation and contrasting narratives on engagement, recall and attitude change with news articles containing data visualization. IEEE Transactions on Visualization and Computer Graphics, 2024. doi: 10 . 1109/TVCG . 2024 . 3355884
  58. E. Sadler-Smith and E. Shefy. The intuitive executive: Understanding and applying ‘gut feel’in decision-making. Academy of Management Perspectives, 18(4):76–91, 2004. doi: 10 . 5465/ame . 2004 . 15268692
  59. Significance of patterns in data visualisations. In Proceedings of the 25th ACM SIGKDD international conference on knowledge discovery & data mining, pp. 1509–1517, 2019. doi: 10 . 1145/3292500 . 3330994
  60. Visual data exploration as a statistical testing procedure: Within-view and between-view multiple comparisons. IEEE Transactions on Visualization and Computer Graphics, 2022. doi: 10 . 1109/TVCG . 2022 . 3175532
  61. Bayesian modeling of human–ai complementarity. Proceedings of the National Academy of Sciences, 119(11):e2111547119, 2022. doi: 10 . 1073/pnas . 2111547119
  62. A. Tversky and D. Kahneman. Judgment under uncertainty: Heuristics and biases: Biases in judgments reveal some heuristics of thinking under uncertainty. science, 185(4157):1124–1131, 1974. doi: 10 . 1126/science . 185 . 4157 . 1124
  63. Practical bayesian model evaluation using leave-one-out cross-validation and waic. Statistics and computing, 27:1413–1432, 2017. doi: 10 . 1007/s11222-016-9696-4
  64. Warning, bias may occur: A proposed approach to detecting cognitive bias in interactive visual analytics. In 2017 ieee conference on visual analytics science and technology (vast), pp. 104–115. IEEE, 2017. doi: 10 . 1109/VAST . 2017 . 8585669
  65. Autods: Towards human-centered automation of data science. In Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems, pp. 1–12, 2021. doi: 10 . 1145/3411764 . 3445526
  66. From human-human collaboration to human-ai collaboration: Designing ai systems that can work together with people. In Extended abstracts of the 2020 CHI Conference on Human Factors in Computing Systems, pp. 1–6, 2020. doi: 10 . 1145/3334480 . 3381069
  67. Human-ai collaboration in data science: Exploring data scientists’ perceptions of automated ai. Proceedings of the ACM on human-computer interaction, 3(CSCW):1–24, 2019. doi: 10 . 1145/3359313
  68. Graphical inference for infovis. IEEE Transactions on Visualization and Computer Graphics, 16(6):973–979, 2010. doi: 10 . 1109/TVCG . 2010 . 161
  69. The rational agent benchmark for data visualization. IEEE Transactions on Visualization and Computer Graphics, 2023. doi: 10 . 1109/TVCG . 2023 . 3326513
  70. Investigating the effect of the multiple comparisons problem in visual analysis. In Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems, pp. 1–12, 2018. doi: 10 . 1145/3173574 . 3174053
  71. Controlling false discoveries during interactive data exploration. In Proceedings of the 2017 ACM International Conference on Management of Data, pp. 527–540, 2017. doi: 10 . 1145/3035918 . 3064019
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Ratanond Koonchanok (4 papers)
  2. Michael E. Papka (25 papers)
  3. Khairi Reda (7 papers)
Youtube Logo Streamline Icon: https://streamlinehq.com