Negotiating the Shared Agency between Humans & AI in the Recommender System (2403.15919v3)
Abstract: Smart recommendation algorithms have revolutionized information dissemination, enhancing efficiency and reshaping content delivery across various domains. However, concerns about user agency have arisen due to the inherent opacity (information asymmetry) and the nature of one-way output (power asymmetry) on algorithms. While both issues have been criticized by scholars via advocating explainable AI (XAI) and human-AI collaborative decision-making (HACD), few research evaluates their integrated effects on users, and few HACD discussions in recommender systems beyond improving and filtering the results. This study proposes an incubating idea as a missing step in HACD that allows users to control the degrees of AI-recommended content. Then, we integrate it with existing XAI to a flow prototype aimed at assessing the enhancement of user agency. We seek to understand how types of agency impact user perception and experience, and bring empirical evidence to refine the guidelines and designs for human-AI interactive systems.
- Cookie disclaimers: Dark patterns and lack of transparency. Computers & Security 136 (2024), 103507.
- Jenna Burrell. 2016. How the machine ‘thinks’: Understanding opacity in machine learning algorithms. Big data & society 3, 1 (2016), 2053951715622512.
- Machine learning interpretability: A survey on methods and metrics. Electronics 8, 8 (2019), 832.
- Tsai-Wei Chen and S Shyam Sundar. 2018. This app would like to use your current location to better serve you: Importance of user assent and system transparency in personalized mobile services. In Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems. ACM, USA, 1–13.
- Lingwei Cheng and Alexandra Chouldechova. 2023. Overcoming Algorithm Aversion: A Comparison between Process and Outcome Control. In Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems. ACM, USA, 1–27.
- Political polarization on twitter. In Proceedings of the International AAAI Conference on Web and Social Media, Vol. 5-1. AAAI, USA, 89–96.
- The effects of transparency on trust in and acceptance of a content-based art recommender. User Modeling and User-adapted Interaction 18 (2008), 455–496.
- Bringing transparency design into practice. In 23rd international conference on intelligent user interfaces. ACM, USA, 211–223.
- Mica R Endsley. 2023. Supporting Human-AI Teams: Transparency, explainability, and situation awareness. Computers in Human Behavior 140 (2023), 107574.
- Towards transparency by design for artificial intelligence. Science and Engineering Ethics 26, 6 (2020), 3333–3361.
- Explaining explanations: An overview of interpretability of machine learning. In 2018 IEEE 5th International Conference on data science and advanced analytics (DSAA). IEEE, IEEE, USA, 80–89.
- Design, development and evaluation of a human-computer trust scale. Behaviour & Information Technology 38, 10 (2019), 1004–1015.
- René F Kizilcec. 2016. How much information? Effects of transparency on trust in an algorithmic interface. In Proceedings of the 2016 CHI Conference on Human Factors in Computing Systems. ACM, USA, 2390–2395.
- Explaining the user experience of recommender systems. User modeling and user-adapted interaction 22 (2012), 441–504.
- Towards a Science of Human-AI Decision Making: An Overview of Design Space in Empirical Human-Subject Studies. In Proceedings of the 2023 ACM Conference on Fairness, Accountability, and Transparency. ACM, USA, 1369–1385.
- Bingjie Liu. 2021. In AI we trust? Effects of agency locus and transparency on uncertainty reduction in human–AI interaction. Journal of Computer-Mediated Communication 26, 6 (2021), 384–402.
- Cookies and web browser design: Toward realizing informed consent online. In Proceedings of the SIGCHI conference on Human factors in computing systems. ACM, USA, 46–52.
- Maria D Molina and S Shyam Sundar. 2022. When AI moderates online content: effects of human collaboration and interactive transparency on user trust. Journal of Computer-Mediated Communication 27, 4 (2022), zmac010.
- Exploring the filter bubble: the effect of using recommender systems on content diversity. In Proceedings of the 23rd International Conference on World Wide Web. ACM, USA, 677–686.
- Frank Pasquale. 2015. The black box society: The secret algorithms that control money and information. Harvard University Press, USA.
- A user-centric evaluation framework for recommender systems. In Proceedings of the fifth ACM conference on Recommender systems. ACM, USA, 157–164.
- Explanations as mechanisms for supporting algorithmic transparency. In Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems. ACM, USA, 1–13.
- Explanation methods in deep learning: Users, values, concerns and challenges. Explainable and interpretable models in computer vision and machine learning (2018), 19–36.
- Getting to know you: learning new user preferences in recommender systems. In Proceedings of the 7th international conference on Intelligent user interfaces. ACM, USA, 127–134.
- ” Why should I trust you?” Explaining the predictions of any classifier. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. ACM, USA, 1135–1144.
- Fred Rowland. 2011. The filter bubble: what the internet is hiding from you. portal: Libraries and the Academy 11, 4 (2011), 1009–1011.
- Intelligent IT Systems in Business Application: Control and Transparency as Means of Building Trust in AI. In Work and AI 2030: Challenges and Strategies for Tomorrow’s Work. Springer, USA, 125–132.
- A meta-analysis of the utility of explainable artificial intelligence in human-AI decision-making. In Proceedings of the 2022 AAAI/ACM Conference on AI, Ethics, and Society. ACM, USA, 617–626.
- A literature review of personalization transparency and control: introducing the transparency–awareness–control Framework. Media and Communication 9, 4 (2021), 120–133.
- Anyuan Shen. 2014. Recommendations as personalized marketing: insights from customer experiences. Journal of Services Marketing 28, 5 (2014), 414–427.
- Fairness and transparency in recommendation: The users’ perspective. In Proceedings of the 29th ACM Conference on User Modeling, Adaptation and Personalization. ACM, USA, 274–279.
- S Shyam Sundar. 2020. Rise of machine agency: A framework for studying the psychology of human-AI interaction (HAII). Journal of Computer-Mediated Communication 25, 1 (2020), 74–88.
- Adrian Weller. 2019. Transparency: motivations and challenges. In Explainable AI: Interpreting, Explaining and Visualizing Deep Learning. Springer, USA, 23–40.
- Christine T Wolf. 2019. Explainability scenarios: towards scenario-based XAI design. In Proceedings of the 24th International Conference on Intelligent User Interfaces. ACM, USA, 252–257.
- Bo Zhang and S Shyam Sundar. 2019. Proactive vs. reactive personalization: Can customization of privacy enhance user experience? International journal of human-computer studies 128 (2019), 86–99.
- PERD: Personalized emoji recommendation with dynamic user preference. In Proceedings of the 45th international ACM SIGIR conference on research and development in information retrieval. ACM, USA, 1922–1926.
- Mengke Wu (1 paper)
- Weizi Liu (3 papers)
- Yanyun Wang (10 papers)
- Mike Yao (1 paper)