Unveiling the Impact of Multi-Modal Interactions on User Engagement: A Comprehensive Evaluation in AI-driven Conversations (2406.15000v1)
Abstract: LLMs have significantly advanced user-bot interactions, enabling more complex and coherent dialogues. However, the prevalent text-only modality might not fully exploit the potential for effective user engagement. This paper explores the impact of multi-modal interactions, which incorporate images and audio alongside text, on user engagement in chatbot conversations. We conduct a comprehensive analysis using a diverse set of chatbots and real-user interaction data, employing metrics such as retention rate and conversation length to evaluate user engagement. Our findings reveal a significant enhancement in user engagement with multi-modal interactions compared to text-only dialogues. Notably, the incorporation of a third modality significantly amplifies engagement beyond the benefits observed with just two modalities. These results suggest that multi-modal interactions optimize cognitive processing and facilitate richer information comprehension. This study underscores the importance of multi-modality in chatbot design, offering valuable insights for creating more engaging and immersive AI communication experiences and informing the broader AI community about the benefits of multi-modal interactions in enhancing user engagement.
- Lichao Zhang (17 papers)
- Jia Yu (54 papers)
- Shuai Zhang (319 papers)
- Long Li (113 papers)
- Yangyang Zhong (4 papers)
- Guanbao Liang (1 paper)
- Yuming Yan (7 papers)
- Qing Ma (6 papers)
- Fangsheng Weng (4 papers)
- Fayu Pan (2 papers)
- Jing Li (621 papers)
- Renjun Xu (28 papers)
- Zhenzhong Lan (56 papers)