User Characteristics in Explainable AI: The Rabbit Hole of Personalization? (2403.00137v1)
Abstract: As AI becomes ubiquitous, the need for Explainable AI (XAI) has become critical for transparency and trust among users. A significant challenge in XAI is catering to diverse users, such as data scientists, domain experts, and end-users. Recent research has started to investigate how users' characteristics impact interactions with and user experience of explanations, with a view to personalizing XAI. However, are we heading down a rabbit hole by focusing on unimportant details? Our research aimed to investigate how user characteristics are related to using, understanding, and trusting an AI system that provides explanations. Our empirical study with 149 participants who interacted with an XAI system that flagged inappropriate comments showed that very few user characteristics mattered; only age and the personality trait openness influenced actual understanding. Our work provides evidence to reorient user-focused XAI research and question the pursuit of personalized XAI based on fine-grained user characteristics.
- Trends and Trajectories for Explainable, Accountable and Intelligible Systems: An HCI Research Agenda. In Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems (CHI ’18). Association for Computing Machinery, New York, NY, USA, 1–18. https://doi.org/10.1145/3173574.3174156
- Amina Adadi and Mohammed Berrada. 2018. Peeking inside the black-box: a survey on explainable artificial intelligence (XAI). IEEE access 6 (2018), 52138–52160. https://doi.org/10.1109/access.2018.2870052 Publisher: IEEE.
- Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI. Information Fusion 58 (June 2020), 82–115. https://doi.org/10.1016/j.inffus.2019.12.012
- Verónica Benet-Martínez and Oliver P. John. 1998. Los Cinco Grandes across cultures and ethnic groups: Multitrait-multimethod analyses of the Big Five in Spanish and English. Journal of Personality and Social Psychology 75 (1998), 729–750. https://doi.org/10.1037/0022-3514.75.3.729
- GenderMag: A Method for Evaluating Software’s Gender Inclusiveness. Interacting with Computers 28, 6 (Nov. 2016), 760–787. https://doi.org/10.1093/iwc/iwv046
- Proxy tasks and subjective measures can be misleading in evaluating explainable AI systems. In Proceedings of the 25th International Conference on Intelligent User Interfaces (IUI ’20). Association for Computing Machinery, New York, NY, USA, 454–464. https://doi.org/10.1145/3377325.3377498
- To Trust or to Think: Cognitive Forcing Functions Can Reduce Overreliance on AI in AI-assisted Decision-making. Proceedings of the ACM on Human-Computer Interaction 5, CSCW1 (April 2021), 188:1–188:21. https://doi.org/10.1145/3449287
- John T Cacioppo and Richard E Petty. 1982. The need for cognition. Journal of personality and social psychology 42, 1 (1982), 116. https://doi.org/10.1037/0022-3514.42.1.116 Publisher: American Psychological Association.
- Impacts of Personal Characteristics on User Trust in Conversational Recommender Systems. In Proceedings of the 2022 CHI Conference on Human Factors in Computing Systems (CHI ’22). Association for Computing Machinery, New York, NY, USA, 1–14. https://doi.org/10.1145/3491102.3517471
- Toxic Comment Classification Challenge. https://kaggle.com/competitions/jigsaw-toxic-comment-classification-challenge
- Toward personalized XAI: A case study in intelligent tutoring systems. Artificial Intelligence 298 (2021), 103503. https://doi.org/10.1016/j.artint.2021.103503
- Do We Really Like Robots that Match our Personality? The Case of Big-Five Traits, Godspeed Scores and Robotic Gestures. In 2018 27th IEEE International Symposium on Robot and Human Interactive Communication (RO-MAN). IEEE, Nanjing, China, 626–631. https://doi.org/10.1109/ROMAN.2018.8525672
- Shaping Gestures to Shape Personalities: The Relationship Between Gesture Parameters, Attributed Personality Traits and Godspeed Scores. In 2018 27th IEEE International Symposium on Robot and Human Interactive Communication (RO-MAN). IEEE, Nanjing, China, 699–704. https://doi.org/10.1109/ROMAN.2018.8525739
- The effects of transparency on trust in and acceptance of a content-based art recommender. User Modeling and User-Adapted Interaction 18, 5 (Nov. 2008), 455–496. https://doi.org/10.1007/s11257-008-9051-3
- A Case for Humans-in-the-Loop: Decisions in the Presence of Erroneous Algorithmic Scores. In Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems (CHI ’20). Association for Computing Machinery, New York, NY, USA, 1–12. https://doi.org/10.1145/3313831.3376638
- BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. https://doi.org/10.48550/arXiv.1810.04805
- A Survey on Personality-Aware Recommendation Systems. Artificial Intelligence Review 55, 3 (March 2022), 2409–2454. https://doi.org/10.1007/s10462-021-10063-7
- Explaining Models: An Empirical Study of How Explanations Impact Fairness Judgment. In Proceedings of the 24th International Conference on Intelligent User Interfaces (IUI ’19). Association for Computing Machinery, New York, NY, USA, 275–285. https://doi.org/10.1145/3301275.3302310
- Finale Doshi-Velez and Been Kim. 2017. Towards A Rigorous Science of Interpretable Machine Learning. (March 2017). https://doi.org/10.48550/arxiv.1702.08608
- The role of trust in automation reliance. International Journal of Human-Computer Studies 58 (2003), 697–718. https://doi.org/10.1016/S1071-5819(03)00038-7
- The Who in Explainable AI: How AI Background Shapes Perceptions of AI Explanations. https://doi.org/10.48550/arXiv.2107.13509
- Upol Ehsan and Mark O. Riedl. 2020. Human-centered Explainable AI: Towards a Reflective Sociotechnical Approach. https://doi.org/10.48550/arXiv.2002.01092
- Operationalizing Human-Centered Perspectives in Explainable AI. In Extended Abstracts of the 2021 CHI Conference on Human Factors in Computing Systems (CHI EA ’21). Association for Computing Machinery, New York, NY, USA, 1–6. https://doi.org/10.1145/3411763.3441342
- A Study on Trust in Black Box Models and Post-hoc Explanations. In 14th International Conference on Soft Computing Models in Industrial and Environmental Applications (SOCO 2019) (Advances in Intelligent Systems and Computing), Francisco Martínez Álvarez, Alicia Troncoso Lora, José António Sáez Muñoz, Héctor Quintián, and Emilio Corchado (Eds.). Springer International Publishing, Cham, 35–46. https://doi.org/10.1007/978-3-030-20055-8_4
- Gender and Participation in Open Source Software Development. Proceedings of the ACM on Human-Computer Interaction 6, CSCW2. https://doi.org/10.1145/3555190
- Nirit Glazer. 2011. Challenges with graph interpretation: a review of the literature. Studies in Science Education 47, 2 (Sept. 2011), 183–210. https://doi.org/10.1080/03057267.2011.605307
- A very brief measure of the Big-Five personality domains. Journal of Research in Personality 37, 6 (Dec. 2003), 504–528. https://doi.org/10.1016/S0092-6566(03)00046-1
- XAI—Explainable artificial intelligence. Science Robotics 4, 37 (Dec. 2019), eaay7120. https://doi.org/10.1126/scirobotics.aay7120
- How and why humans trust: A meta-analysis and elaborated model. Frontiers in Psychology 14 (March 2023), 1081086. https://doi.org/10.3389/fpsyg.2023.1081086
- Laura Hanu and Unitary team. 2020. Detoxify. https://github.com/unitaryai/detoxify
- Situated Data, Situated Systems: A Methodology to Engage with Power Relations in Natural Language Processing Research. In Proceedings of the Second Workshop on Gender Bias in Natural Language Processing, Marta R. Costa-jussà, Christian Hardmeier, Will Radford, and Kellie Webster (Eds.). Association for Computational Linguistics, Barcelona, Spain (Online), 107–124. https://aclanthology.org/2020.gebnlp-1.10
- Metrics for Explainable AI: Challenges and Prospects. http://arxiv.org/abs/1812.04608
- Who needs explanation and when? Juggling explainable AI and user epistemic uncertainty. International Journal of Human-Computer Studies 165 (2022), 102839. https://doi.org/10.1016/j.ijhcs.2022.102839
- The big five inventory—versions 4a and 54. Berkeley, CA: University of California, Berkeley, Institute of Personality.
- Paradigm shift to the integrative Big Five trait taxonomy: History, measurement, and conceptual issues. In Handbook of personality: Theory and research, 3rd ed. The Guilford Press, New York, NY, USA, 114–158.
- The Roles of Personality Traits, AI Anxiety, and Demographic Factors in Attitudes toward Artificial Intelligence. International Journal of Human–Computer Interaction 0, 0 (Dec. 2022), 1–18. https://doi.org/10.1080/10447318.2022.2151730
- The Age of AI: And Our Human Future. John Murray Press.
- Explaining the user experience of recommender systems. User Modeling and User-Adapted Interaction 22, 4 (Oct. 2012), 441–504. https://doi.org/10.1007/s11257-011-9118-4
- Principles of Explanatory Debugging to Personalize Interactive Machine Learning. In Proceedings of the 20th International Conference on Intelligent User Interfaces (IUI ’15). Association for Computing Machinery, New York, NY, USA, 126–137. https://doi.org/10.1145/2678025.2701399
- Tell me more? the effects of mental model soundness on personalizing an intelligent agent. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI ’12). Association for Computing Machinery, New York, NY, USA, 1–10. https://doi.org/10.1145/2207676.2207678
- Explanatory Debugging: Supporting End-User Debugging of Machine-Learned Programs. In 2010 IEEE Symposium on Visual Languages and Human-Centric Computing. IEEE, NW Washington, DC, USA, 41–48. https://doi.org/10.1109/VLHCC.2010.15
- Too much, too little, or just right? Ways explanations impact end users’ mental models. In 2013 IEEE Symposium on Visual Languages and Human Centric Computing. IEEE, San Jose, CA, USA, 3–10. https://doi.org/10.1109/VLHCC.2013.6645235
- Why-oriented end-user debugging of naive Bayes text classification. ACM Transactions on Interactive Intelligent Systems 1, 1 (Oct. 2011), 2:1–2:31. https://doi.org/10.1145/2030365.2030367
- Fixing the program my computer learned: barriers for end users, challenges for the machine. In Proceedings of the 14th international conference on Intelligent user interfaces (IUI ’09). Association for Computing Machinery, New York, NY, USA, 187–196. https://doi.org/10.1145/1502650.1502678
- Effects of Explainable Artificial Intelligence on trust and human behavior in a high-risk decision task. Computers in Human Behavior 139 (2023), 107539. https://doi.org/10.1016/j.chb.2022.107539
- User Trust in Recommendation Systems: A comparison of Content-Based, Collaborative and Demographic Filtering. In Proceedings of the 2022 CHI Conference on Human Factors in Computing Systems (CHI ’22). Association for Computing Machinery, New York, NY, USA, 1–14. https://doi.org/10.1145/3491102.3501936
- Introduction to Explainable AI. In Extended Abstracts of the 2021 CHI Conference on Human Factors in Computing Systems (CHI EA ’21). Association for Computing Machinery, New York, NY, USA, 1–3. https://doi.org/10.1145/3411763.3445016
- Q. Vera Liao and Kush R. Varshney. 2022. Human-Centered Explainable AI (XAI): From Algorithms to User Experiences. https://doi.org/10.48550/arXiv.2110.10790
- Brian Y. Lim and Anind K. Dey. 2009. Assessing demand for intelligibility in context-aware applications. In Proceedings of the 11th international conference on Ubiquitous computing (UbiComp ’09). Association for Computing Machinery, New York, NY, USA, 195–204. https://doi.org/10.1145/1620545.1620576
- Why and why not explanations improve the intelligibility of context-aware intelligent systems. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI ’09). Association for Computing Machinery, New York, NY, USA, 2119–2128. https://doi.org/10.1145/1518701.1519023
- Scott Lundberg and Su-In Lee. 2017. A Unified Approach to Interpreting Model Predictions. https://doi.org/10.48550/arXiv.1705.07874
- How Does Domain Expertise Affect Users’ Search Interaction and Outcome in Exploratory Search? ACM Transactions on Information Systems 36, 4 (July 2018), 42:1–42:30. https://doi.org/10.1145/3223045
- Jim Mcbeath. 2011. Levels of Expertise. http://jim-mcbeath.blogspot.com/2011/12/levels-of-expertise.html Publication Title: Jim McBeath.
- R. R. McCrae and P. T. Costa. 1987. Validation of the five-factor model of personality across instruments and observers. Journal of Personality and Social Psychology 52, 1 (Jan. 1987), 81–90. https://doi.org/10.1037//0022-3514.52.1.81
- Jeremy Miles. 2014. Tolerance and variance inflation factor. Wiley statsref: statistics reference online (2014). https://doi.org/10.1002/9781118445112.stat06593 Publisher: Wiley Online Library.
- To explain or not to explain: the effects of personal characteristics when explaining music recommendations. In Proceedings of the 24th International Conference on Intelligent User Interfaces (IUI ’19). Association for Computing Machinery, New York, NY, USA, 397–407. https://doi.org/10.1145/3301275.3302313
- What’s in a User? Towards Personalising Transparency for Music Recommender Interfaces. In Proceedings of the 28th ACM Conference on User Modeling, Adaptation and Personalization (UMAP ’20). Association for Computing Machinery, New York, NY, USA, 173–182. https://doi.org/10.1145/3340631.3394844
- Tim Miller. 2019. Explanation in artificial intelligence: Insights from the social sciences. Artificial Intelligence 267 (Feb. 2019), 1–38. https://doi.org/10.1016/j.artint.2018.07.007
- Explainable AI: Beware of Inmates Running the Asylum Or: How I Learnt to Stop Worrying and Love the Social and Behavioural Sciences. http://arxiv.org/abs/1712.00547
- D. A. Norman. 1987. Some observations on mental models. In Human-computer interaction: a multidisciplinary approach. Morgan Kaufmann Publishers Inc., San Francisco, CA, USA, 241–244. https://dl.acm.org/doi/10.5555/58076.58097
- Judea Pearl. 2009. Causality. Cambridge university press. https://doi.org/10.1017/CBO9780511803161
- Opening the Black Box: The Promise and Limitations of Explainable Machine Learning in Cardiology. Canadian Journal of Cardiology 38, 2 (Feb. 2022), 204–213. https://doi.org/10.1016/j.cjca.2021.09.004
- Mashup-based information retrieval for domain experts. In Proceedings of the 18th ACM conference on Information and knowledge management (CIKM ’09). Association for Computing Machinery, New York, NY, USA, 711–720. https://doi.org/10.1145/1645953.1646044
- Evaluating Explainable AI (XAI) in Terms of User Gender and Educational Background. In Artificial Intelligence in HCI (Lecture Notes in Computer Science), Helmut Degen and Stavroula Ntoa (Eds.). Springer Nature Switzerland, Cham, 286–304. https://doi.org/10.1007/978-3-031-35891-3_18
- ”Why Should I Trust You?”: Explaining the Predictions of Any Classifier. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD ’16). Association for Computing Machinery, New York, NY, USA, 1135–1144. https://doi.org/10.1145/2939672.2939778
- Marco Tulio Correia Ribeiro. 2023. lime. https://github.com/marcotcr/lime
- Lloyd S. Shapley. 1951. Notes on the N-Person Game — II: The Value of an N-Person Game. Technical Report. RAND Corporation. https://www.rand.org/pubs/research_memoranda/RM0670.html
- Kacper Sokol and Peter Flach. 2018. Glass-box: explaining AI decisions with counterfactual statements through conversation with a voice-enabled virtual assistant. In Proceedings of the 27th International Joint Conference on Artificial Intelligence (IJCAI’18). AAAI Press, 5868–5870. https://doi.org/10.24963/ijcai.2018/865 Place: Stockholm, Sweden.
- Age differences in personality traits from 10 to 65: Big Five domains and facets in a large cross-sectional sample. Journal of Personality and Social Psychology 100 (2011), 330–348. https://doi.org/10.1037/a0021717
- On domain expertise-based roles in collaborative information retrieval. Information Processing & Management 50, 5 (Sept. 2014), 752–774. https://doi.org/10.1016/j.ipm.2014.04.002
- Interacting meaningfully with machine learning systems: Three experiments. International Journal of Human-Computer Studies 67, 8 (Aug. 2009), 639–662. https://doi.org/10.1016/j.ijhcs.2009.03.004
- Integrating rich user feedback into intelligent user interfaces. In Proceedings of the 13th international conference on Intelligent user interfaces (IUI ’08). Association for Computing Machinery, New York, NY, USA, 50–59. https://doi.org/10.1145/1378773.1378781
- Axiomatic Attribution for Deep Networks. In Proceedings of the 34th International Conference on Machine Learning (Proceedings of Machine Learning Research, Vol. 70), Doina Precup and Yee Whye Teh (Eds.). PMLR, 3319–3328. https://proceedings.mlr.press/v70/sundararajan17a.html
- Visual, textual or hybrid: the effect of user expertise on different explanations. In 26th International Conference on Intelligent User Interfaces (IUI ’21). Association for Computing Machinery, New York, NY, USA, 109–119. https://doi.org/10.1145/3397481.3450662
- Human-Centered Responsible Artificial Intelligence: Current & Future Trends. In Extended Abstracts of the 2023 CHI Conference on Human Factors in Computing Systems. 1–4. https://doi.org/10.1145/3544549.3583178
- Personality-based tailored explainable recommendation for trustworthy smart learning system in the age of artificial intelligence. Smart Learning Environments 10, 1 (2023), 65. https://doi.org/10.1186/s40561-023-00282-6 Publisher: Springer.
- Adapting Recommendation Diversity to Openness to Experience: A Study of Human Behaviour. In User Modeling, Adaptation, and Personalization (Lecture Notes in Computer Science), Sandra Carberry, Stephan Weibelzahl, Alessandro Micarelli, and Giovanni Semeraro (Eds.). Springer, Berlin, Heidelberg, 190–202. https://doi.org/10.1007/978-3-642-38844-6_16
- Nava Tintarev and Judith Masthoff. 2007. A Survey of Explanations in Recommender Systems. In 2007 IEEE 23rd International Conference on Data Engineering Workshop. IEEE, Istanbul, Turkey, 801–810. https://doi.org/10.1109/ICDEW.2007.4401070
- Marko Tkalcic and Li Chen. 2015. Personality and Recommender Systems. In Recommender Systems Handbook, Francesco Ricci, Lior Rokach, and Bracha Shapira (Eds.). Springer US, Boston, MA, 715–739. https://doi.org/10.1007/978-1-4899-7637-6_21
- User Needs for Explanations of Recommendations: In-depth Analyses of the Role of Item Domain and Personal Characteristics. In Proceedings of the 31st ACM Conference on User Modeling, Adaptation and Personalization (UMAP ’23). Association for Computing Machinery, New York, NY, USA, 54–65. https://doi.org/10.1145/3565472.3592950
- Characterizing the influence of domain expertise on web search behavior. In Proceedings of the Second ACM International Conference on Web Search and Data Mining (WSDM ’09). Association for Computing Machinery, New York, NY, USA, 132–141. https://doi.org/10.1145/1498759.1498819
- Concept-based document readability in domain specific information retrieval. In Proceedings of the 15th ACM international conference on Information and knowledge management (CIKM ’06). Association for Computing Machinery, New York, NY, USA, 540–549. https://doi.org/10.1145/1183614.1183692
- Robert Nimmo (1 paper)
- Marios Constantinides (35 papers)
- Ke Zhou (48 papers)
- Daniele Quercia (77 papers)
- Simone Stumpf (16 papers)