CloChat: Understanding How People Customize, Interact, and Experience Personas in Large Language Models (2402.15265v1)
Abstract: LLMs have facilitated significant strides in generating conversational agents, enabling seamless, contextually relevant dialogues across diverse topics. However, the existing LLM-driven conversational agents have fixed personalities and functionalities, limiting their adaptability to individual user needs. Creating personalized agent personas with distinct expertise or traits can address this issue. Nonetheless, we lack knowledge of how people customize and interact with agent personas. In this research, we investigated how users customize agent personas and their impact on interaction quality, diversity, and dynamics. To this end, we developed CloChat, an interface supporting easy and accurate customization of agent personas in LLMs. We conducted a study comparing how participants interact with CloChat and ChatGPT. The results indicate that participants formed emotional bonds with the customized agents, engaged in more dynamic dialogues, and showed interest in sustaining interactions. These findings contribute to design implications for future systems with conversational agents using LLMs.
- [n. d.]. F/awesome-CHATGPT-prompts. https://github.com/f/awesome-chatgpt-prompts
- [n. d.]. Google Bard. https://bard.google.com/
- [n. d.]. Introducing ChatGPT. https://openai.com/blog/chatgpt/
- [n. d.]. Introducing GPTs. https://openai.com/blog/introducing-gpts
- KITLM: Domain-Specific Knowledge InTegration into Language Models for Question Answering. arXiv:2308.03638 [cs.CL]
- Open AI. 2023. Introducing GPTs. https://openai.com/blog/introducing-gpts
- Fine-Tuning GPT-3 for Russian Text Summarization. In Data Science and Intelligent Systems, Radek Silhavy, Petr Silhavy, and Zdenka Prokopova (Eds.). Springer International Publishing, Cham, 748–757.
- Toxicity in chatGPT: Analyzing persona-assigned language models. arXiv preprint arXiv: 2304.05335 (2023). https://doi.org/10.48550/arXiv.2304.05335
- An empirical investigation of the influence of persona with personality traits on conceptual design. Journal of Systems and Software 134 (2017), 324–339. https://doi.org/10.1016/j.jss.2017.09.020
- A multitask, multilingual, multimodal evaluation of ChatGPT on reasoning, hallucination, and interactivity. arXiv. https://doi.org/10.48550/arXiv.2302.04023
- The Chatbot Usability Scale: the design and pilot of a usability scale for interaction with AI-based conversational agents. Personal and Ubiquitous Computing 26 (2022), 95–119. https://doi.org/10.1007/s00779-021-01582-9
- Petter Bae Brandtzaeg and Asbjørn Følstad. 2018. Chatbots: changing user needs and motivations. interactions 25, 5 (2018), 38–43. https://doi.org/10.1145/3236669
- At your service: Designing voice assistant personalities to improve automotive user interfaces. In Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems. 1–11. https://doi.org/10.1145/3290605.3300270
- Virginia Braun and Victoria Clarke. 2012. Thematic analysis. American Psychological Association.
- Sparks of artificial general intelligence: Early experiments with gpt-4. arXiv preprint arXiv:2303.12712 (2023). https://doi.org/10.48550/arXiv.2303.12712
- Personas: From Theory to Practices (NordiCHI ’08). Association for Computing Machinery, New York, NY, USA, 439–442. https://doi.org/10.1145/1463160.1463214
- Personas: from theory to practices. In Proceedings of the 5th Nordic conference on Human-computer interaction: building bridges. 439–442. https://doi.org/10.1145/1463160.1463214
- Ana Paula Chaves and Marco Aurelio Gerosa. 2021. How Should My Chatbot Interact? A Survey on Social Characteristics in Human–Chatbot Interaction Design. International Journal of Human–Computer Interaction 37, 8 (2021), 729–758. https://doi.org/10.1080/10447318.2020.1841438
- Salesbot: Transitioning from chit-chat to task-oriented dialogues. arXiv preprint arXiv:2204.10591 (2022). https://doi.org/10.48550/arXiv.2204.10591
- ” What can i help you with?” infrequent users’ experiences of intelligent personal assistants. In Proceedings of the 19th international conference on human-computer interaction with mobile devices and services. 1–12. https://doi.org/10.1145/3098279.3098539
- Carolyn E Cutrona and Julie A Suhr. 1992. Controllability of stressful events and satisfaction with spouse support behaviors. Communication research 19, 2 (1992), 154–174. https://doi.org/10.1177/009365092019002002
- Chatbot personality and customer satisfaction. Info Support Research (2018). https://research.infosupport.com/wp-content/uploads/Chatbot-Personality-and-Customer-Satisfaction-Bachelor-Thesis-Information-Sciences-Hayco-de-Haan.pdf
- Toxicity in chatgpt: Analyzing persona-assigned language models. arXiv preprint arXiv:2304.05335 (2023).
- Wizard of wikipedia: Knowledge-powered conversational agents. arXiv preprint arXiv:1811.01241 (2018). https://doi.org/10.48550/arXiv.1811.01241
- Parallel Prototyping Leads to Better Design Results, More Divergence, and Increased Self-Efficacy. ACM Trans. Comput.-Hum. Interact. 17, 4, Article 18 (dec 2011), 24 pages. https://doi.org/10.1145/1879831.1879836
- Human conversational behavior. Human nature 8 (1997), 231–246. https://doi.org/10.1007/BF02912493
- Nature Editorial. 2023. ChatGPT is a black box: how AI research can break it open. https://www.nature.com/articles/d41586-023-02366-2
- Towards automated dialog personalization using mbti personality indicators. In Proc. Interspeech. 1968–1972. https://doi.org/10.21437/Interspeech.2022-376
- Delivering cognitive behavior therapy to young adults with symptoms of depression and anxiety using a fully automated conversational agent (Woebot): a randomized controlled trial. JMIR mental health 4, 2 (2017), e7785. https://doi.org/10.2196/mental.7785
- Social support in a virtual community: analysis of a clinic-affiliated online support group for persons living with HIV/AIDS. AIDS and Behavior 21 (2017), 3087–3099. https://doi.org/10.1007/s10461-016-1587-3
- Luciano Floridi and Massimo Chiriatti. 2020. GPT-3: Its nature, scope, limits, and consequences. Minds and Machines 30 (2020), 681–694. https://doi.org/10.1007/s11023-020-09548-1
- Neural approaches to conversational AI. In The 41st international ACM SIGIR conference on research & development in information retrieval. 1371–1374. https://doi.org/10.1145/3209978.3210183
- A knowledge-grounded neural conversation model. In Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 32. https://doi.org/10.1609/aaai.v32i1.11977
- Anand Gokul. 2023. LLMs and AI: Understanding Its Reach and Impact. (2023).
- Generative language models and automated influence operations: Emerging threats and potential mitigations. arXiv preprint arXiv:2301.04246 (2023). https://doi.org/10.48550/arXiv.2301.04246
- Understanding the user experience of customer service chatbots: An experimental study of chatbot interaction design. International Journal of Human-Computer Studies 161 (2022), 102788. https://doi.org/10.1016/j.ijhcs.2022.102788
- hu.ma.ne. 2023. Ai Pin Overview. https://hu.ma.ne/aipin
- Linguistic Features to Consider When Applying Persona of the Real Person to the Text-Based Agent (MobileHCI ’20). Association for Computing Machinery, New York, NY, USA, Article 23, 4 pages. https://doi.org/10.1145/3406324.3410723
- Measuring and Explaining the Inter-Cluster Reliability of Multidimensional Projections. IEEE Transactions on Visualization and Computer Graphics 28, 1 (2022), 551–561. https://doi.org/10.1109/TVCG.2021.3114833
- Uniform Manifold Approximation with Two-phase Optimization. In 2022 IEEE Visualization and Visual Analytics (VIS). IEEE, 80–84. https://doi.org/10.1109/VIS54862.2022.00025
- CLAMS: A Cluster Ambiguity Measure for Estimating Perceptual Variability in Visual Clustering. arXiv preprint arXiv:2308.00284 (2023). https://doi.org/10.48550/arXiv.2308.00284
- Personallm: Investigating the ability of gpt-3.5 to express personality traits and gender differences. arXiv preprint arXiv:2305.02547 (2023). https://doi.org/10.48550/arXiv.2305.02547
- Marius Kaminskas and Derek Bridge. 2016. Diversity, Serendipity, Novelty, and Coverage: A Survey and Empirical Analysis of Beyond-Accuracy Objectives in Recommender Systems. ACM Trans. Interact. Intell. Syst. 7, 1, Article 2 (dec 2016), 42 pages. https://doi.org/10.1145/2926720
- Statistical Practices of Educational Researchers: An Analysis of their ANOVA, MANOVA, and ANCOVA Analyses. Review of Educational Research 68, 3 (1998), 350–386. https://doi.org/10.3102/00346543068003350
- Explaining the user experience of recommender systems. User modeling and user-adapted interaction 22 (2012), 441–504. https://doi.org/10.1007/s11257-011-9118-4
- A Baki Kocaballi. 2023. Conversational ai-powered design: Chatgpt as designer, user, and product. arXiv preprint arXiv:2302.07406 (2023). https://doi.org/10.48550/arXiv.2302.07406
- ” What does your Agent look like?” A Drawing Study to Understand Users’ Perceived Persona of Conversational Agent. In Extended abstracts of the 2019 CHI conference on human factors in computing systems. 1–6. https://doi.org/10.1145/3290607.3312796
- Nadine Lessio and Alexis Morris. 2020. Toward Design Archetypes for Conversational Agent Personality. In 2020 IEEE International Conference on Systems, Man, and Cybernetics (SMC). IEEE, 3221–3228. https://doi.org/10.1109/SMC42975.2020.9283254
- A diversity-promoting objective function for neural conversation models. arXiv preprint arXiv:1510.03055 (2015). https://doi.org/10.48550/arXiv.1510.03055
- You Truly Understand What I Need: Intellectual and Friendly Dialogue Agents grounding Knowledge and Persona. arXiv preprint arXiv:2301.02401 (2023). https://doi.org/10.48550/arXiv.2301.02401
- Li Liu and Vincent G Duffy. 2023. Exploring the Future Development of Artificial Intelligence (AI) Applications in Chatbots: A Bibliometric Analysis. International Journal of Social Robotics 15, 5 (2023), 703–716. https://doi.org/10.1007/s12369-022-00956-0
- Physician voice characteristics and patient satisfaction in online health consultation. Information & Management 57, 5 (2020), 103233. https://doi.org/10.1016/j.im.2019.103233
- Training millions of personalized dialogue agents. arXiv preprint arXiv:1809.01984 (2018). https://doi.org/10.48550/arXiv.1809.01984
- Robert R McCrae and Oliver P John. 1992. An introduction to the five-factor model and its applications. Journal of personality 60, 2 (1992), 175–215. https://doi.org/10.1111/j.1467-6494.1992.tb00970.x
- Sara Moussawi and Raquel Benbunan-Fich. 2021. The effect of voice and humour on users’ perceptions of personal intelligent agents. Behaviour & Information Technology 40, 15 (2021), 1603–1626. https://doi.org/10.1080/0144929X.2020.1772368
- How perceptions of intelligence and anthropomorphism affect adoption of personal intelligent agents. Electronic Markets 31 (2021), 343–364. https://doi.org/10.1007/s12525-020-00411-w
- A review of current trends in the development of chatbot systems. In 2020 6th International conference on advanced computing and communication systems (ICACCS). IEEE, 706–710. https://doi.org/10.1109/ICACCS48705.2020.9074420
- Ha Nguyen. 2022. Examining Teenagers’ Perceptions of Conversational Agents in Learning Settings. In Proceedings of the 21st Annual ACM Interaction Design and Children Conference (Braga, Portugal) (IDC ’22). Association for Computing Machinery, New York, NY, USA, 374–381. https://doi.org/10.1145/3501712.3529740
- S Nithuna and CA Laseena. 2020. Review on implementation techniques of chatbot. In 2020 International Conference on Communication and Signal Processing (ICCSP). IEEE, 0157–0161. https://doi.org/10.1109/ICCSP48568.2020.9182168
- Capabilities of gpt-4 on medical challenge problems. arXiv preprint arXiv:2303.13375 (2023). https://doi.org/10.48550/arXiv.2303.13375
- Inc Open AI. 2023. Introducing chatgpt. https://openai.com/blog/chatgpt/
- OpenAI. 2023. GPT-4 Technical Report. https://doi.org/10.48550/arXiv.2303.08774 arXiv:2303.08774 [cs.CL]
- Keyu Pan and Yawen Zeng. 2023. Do LLMs Possess a Personality? Making the MBTI Test an Amazing Evaluation for Large Language Models. arXiv preprint arXiv:2307.16180 (2023). https://doi.org/10.48550/arXiv.2307.16180
- Towards understanding human similarity perception in the analysis of large sets of scatter plots. In Proceedings of the 2016 CHI Conference on Human Factors in Computing Systems. 3659–3669. https://doi.org/10.1145/2858036.2858155
- Generative agents: Interactive simulacra of human behavior. arXiv preprint arXiv:2304.03442 (2023). https://doi.org/10.48550/arXiv.2304.03442
- Alisha Pradhan and Amanda Lazar. 2021a. Hey Google, Do You Have a Personality? Designing Personality and Personas for Conversational Agents (CUI ’21). Association for Computing Machinery, New York, NY, USA, Article 12, 4 pages. https://doi.org/10.1145/3469595.3469607
- Alisha Pradhan and Amanda Lazar. 2021b. Hey Google, do you have a personality? Designing personality and personas for conversational agents. In Proceedings of the 3rd Conference on Conversational User Interfaces. 1–4. https://doi.org/10.1145/3469595.3469607
- ” Alexa is my new BFF” social roles, user satisfaction, and personification of the Amazon Echo. In Proceedings of the 2017 CHI conference extended abstracts on human factors in computing systems. 2853–2859. https://doi.org/10.1145/3027063.3053246
- Hierarchical text-conditional image generation with clip latents. arXiv preprint arXiv:2204.06125 1, 2 (2022), 3. https://doi.org/10.48550/arXiv.2204.06125
- Directed Diversity: Leveraging Language Embedding Distances for Collective Creativity in Crowd Ideation. In Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems (Yokohama, Japan) (CHI ’21). Association for Computing Machinery, New York, NY, USA, Article 393, 35 pages. https://doi.org/10.1145/3411764.3445782
- Personality traits in large language models. arXiv preprint arXiv:2307.00184 (2023). https://doi.org/10.48550/arXiv.2307.00184
- The Effect of Experience on Persona Perceptions. In Extended Abstracts of the 2020 CHI Conference on Human Factors in Computing Systems (Honolulu, HI, USA) (CHI EA ’20). Association for Computing Machinery, New York, NY, USA, 1–9. https://doi.org/10.1145/3334480.3382786
- Persona Perception Scale: Developing and Validating an Instrument for Human-Like Representations of Data. In Extended Abstracts of the 2018 CHI Conference on Human Factors in Computing Systems (Montreal QC, Canada) (CHI EA ’18). Association for Computing Machinery, New York, NY, USA, 1–6. https://doi.org/10.1145/3170427.3188461
- Persona Transparency: Analyzing the Impact of Explanations on Perceptions of Data-Driven Personas. International Journal of Human–Computer Interaction 36, 8 (2020), 788–800. https://doi.org/10.1080/10447318.2019.1688946
- Dale Schuurmans. 2023. Memory Augmented Large Language Models are Computationally Universal. arXiv:2301.04589 [cs.CL]
- Revealing persona biases in dialogue systems. arXiv preprint arXiv:2104.08728 (2021). https://doi.org/10.48550/arXiv.2104.08728
- Towards controllable biases in language generation. arXiv preprint arXiv:2005.00268 (2020). https://doi.org/10.48550/arXiv.2005.00268
- Practical bayesian optimization of machine learning algorithms. Advances in neural information processing systems 25 (2012). https://doi.org/10.48550/arXiv.1206.2944
- Viriya Taecharungroj. 2023. “What Can ChatGPT Do?” Analyzing Early Reactions to the Innovative AI Chatbot on Twitter. Big Data and Cognitive Computing 7, 1 (2023), 35. https://doi.org/10.3390/bdcc7010035
- Deborah Tannen. 1984. Conversational Style: Analyzing Talk Among Friends. Vol. 61. 188 pages. https://doi.org/10.2307/414501
- John W. Tukey. 1949. Comparing Individual Means in the Analysis of Variance. Biometrics 5, 2 (1949), 99–114. http://www.jstor.org/stable/3001913
- Stanford University. 2023. Dialogue distillery: Crafting interpolable, interpretable, and introspectable dialogue from LLMs. In Alexa Prize SocialBot Grand Challenge 5 Proceedings. https://www.amazon.science/alexa-prize/proceedings/chirpy-cardinal-dialogue-distillery-crafting-interpolable-interpretable-and-introspectable-dialogue-from-llms
- Aleksandra Urman and Mykola Makhortykh. 2023. The Silence of the LLMs: Cross-Lingual Analysis of Political Bias and False Information Prevalence in ChatGPT, Google Bard, and Bing Chat. (2023).
- Eliciting and analysing users’ envisioned dialogues with perfect voice assistants. In Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems. 1–15. https://doi.org/10.1145/3411764.3445536
- Developing a Personality Model for Speech-Based Conversational Agents Using the Psycholexical Approach (CHI ’20). Association for Computing Machinery, New York, NY, USA, 1–14. https://doi.org/10.1145/3313831.3376210
- User perceptions of extraversion in chatbots after repeated use. In Proceedings of the 2022 CHI Conference on Human Factors in Computing Systems. 1–18. https://doi.org/10.1145/3491102.3502058
- Jieyu Wang and Anita Komlodi. 2012. Children’s Formal and Informal Definition of Technology. In Proceedings of the 2012 IConference (Toronto, Ontario, Canada) (iConference ’12). Association for Computing Machinery, New York, NY, USA, 587–588. https://doi.org/10.1145/2132176.2132299
- Persuasion for good: Towards a personalized persuasive dialogue system for social good. arXiv preprint arXiv:1906.06725 (2019). https://doi.org/10.48550/arXiv.1906.06725
- Philip Weber and Thomas Ludwig. 2020. (Non-) Interacting with conversational agents: perceptions and motivations of using chatbots and voice assistants. In Proceedings of Mensch und Computer 2020. 321–331. https://doi.org/10.1145/3404983.3405513
- Ethical and social risks of harm from language models. arXiv preprint arXiv:2112.04359 (2021). https://doi.org/10.48550/arXiv.2112.04359
- A prompt pattern catalog to enhance prompt engineering with chatgpt. arXiv preprint arXiv:2302.11382 (2023). https://doi.org/10.48550/arXiv.2302.11382
- Chatbot for Health Care and Oncology Applications Using Artificial Intelligence and Machine Learning: Systematic Review. JMIR Cancer 7, 4 (29 Nov 2021), e27850. https://doi.org/10.2196/27850
- Generating stylistic and personalized dialogues for virtual agents in narratives. In Proceedings of the 2023 International Conference on Autonomous Agents and Multiagent Systems. 737–746. https://doi.org/10.5555/3545946.3598706
- Shanshan Yang and Chris Evans. 2019. Opportunities and challenges in using AI chatbots in higher education. In Proceedings of the 2019 3rd International Conference on Education and E-Learning. 79–83. https://doi.org/10.1145/3371647.3371659
- Llm lies: Hallucinations are not bugs, but features as adversarial examples. arXiv preprint arXiv:2310.01469 (2023).
- GPT-4 outperforms ChatGPT in answering non-English questions related to cirrhosis. medRxiv (2023), 2023–05. https://doi.org/10.1101/2023.05.04.23289482
- User engagement study with virtual agents under different cultural contexts. In Intelligent Virtual Agents: 16th International Conference, IVA 2016, Los Angeles, CA, USA, September 20–23, 2016, Proceedings 16. Springer, 364–368. https://doi.org/10.1007/978-3-319-47665-0_34
- Personalizing Dialogue Agents: I have a dog, do you have pets too? https://doi.org/10.48550/arXiv.1801.07243 arXiv:1801.07243 [cs.AI]
- Large language models are human-level prompt engineers. arXiv preprint arXiv:2211.01910 (2022). https://doi.org/10.48550/arXiv.2211.01910
- Fine-tuning language models from human preferences. arXiv preprint arXiv:1909.08593 (2019). https://doi.org/10.48550/arXiv.1909.08593
- Juhye Ha (1 paper)
- Hyeon Jeon (26 papers)
- DaEun Han (1 paper)
- Jinwook Seo (30 papers)
- Changhoon Oh (4 papers)