Experimental Interface for Multimodal and Large Language Model Based Explanations of Educational Recommender Systems (2402.07910v1)
Abstract: In the age of AI, providing learners with suitable and sufficient explanations of AI-based recommendation algorithm's output becomes essential to enable them to make an informed decision about it. However, the rapid development of AI approaches for educational recommendations and their explainability is not accompanied by an equal level of evidence-based experimentation to evaluate the learning effect of those explanations. To address this issue, we propose an experimental web-based tool for evaluating multimodal and LLM based explainability approaches. Our tool provides a comprehensive set of modular, interactive, and customizable explainability elements, which researchers and educators can utilize to study the role of individual and hybrid explainability methods. We design a two-stage evaluation of the proposed tool, with learners and with educators. Our preliminary results from the first stage show high acceptance of the tool's components, user-friendliness, and an induced motivation to use the explanations for exploring more information about the recommendation.
- Explaining Recommendations in E-Learning: Effects on Adolescents’ Trust, in: 27th International Conference on Intelligent User Interfaces, ACM, Helsinki Finland, 2022, pp. 93–105. URL: https://dl.acm.org/doi/10.1145/3490099.3511140. doi:10.1145/3490099.3511140.
- A Model of Social Explanations for a Conversational Movie Recommendation System, in: Proceedings of the 7th International Conference on Human-Agent Interaction, ACM, Kyoto Japan, 2019, pp. 135–143. URL: https://dl.acm.org/doi/10.1145/3349537.3351899. doi:10.1145/3349537.3351899.
- C.-H. Tsai, P. Brusilovsky, Evaluating Visual Explanations for Similarity-Based Recommendations: User Perception and Performance, in: Proceedings of the 27th ACM Conference on User Modeling, Adaptation and Personalization, ACM, Larnaca Cyprus, 2019, pp. 22–30. URL: https://dl.acm.org/doi/10.1145/3320435.3320465. doi:10.1145/3320435.3320465.