Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
60 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

PersLLM: A Personified Training Approach for Large Language Models (2407.12393v4)

Published 17 Jul 2024 in cs.CL, cs.AI, and cs.CY

Abstract: LLMs exhibit aspects of human-level intelligence that catalyze their application as human-like agents in domains such as social simulations, human-machine interactions, and collaborative multi-agent systems. However, the absence of distinct personalities, such as displaying ingratiating behaviors, inconsistent opinions, and uniform response patterns, diminish LLMs utility in practical applications. Addressing this, the development of personality traits in LLMs emerges as a crucial area of research to unlock their latent potential. Existing methods to personify LLMs generally involve strategies like employing stylized training data for instruction tuning or using prompt engineering to simulate different personalities. These methods only capture superficial linguistic styles instead of the core of personalities and are therefore not stable. In this study, we propose PersLLM, integrating psychology-grounded principles of personality: social practice, consistency, and dynamic development, into a comprehensive training methodology. We incorporate personality traits directly into the model parameters, enhancing the model's resistance to induction, promoting consistency, and supporting the dynamic evolution of personality. Single-agent evaluation validates our method's superiority, as it produces responses more aligned with reference personalities compared to other approaches. Case studies for multi-agent communication highlight its benefits in enhancing opinion consistency within individual agents and fostering collaborative creativity among multiple agents in dialogue contexts, potentially benefiting human simulation and multi-agent cooperation. Additionally, human-agent interaction evaluations indicate that our personified models significantly enhance interactive experiences, underscoring the practical implications of our research.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (43)
  1. Gpt-4 technical report. arXiv preprint arXiv:2303.08774, 2023.
  2. Language models are few-shot learners. Advances in neural information processing systems, 33:1877–1901, 2020.
  3. Charles S Carver. Perspectives on personality (7th Edition). Pearson, 2011.
  4. Benchmarking large language models in retrieval-augmented generation. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 38, pages 17754–17762, 2024.
  5. Xueyong Chen. Lotus Lantern Poetry and Dreams by Lin Huiyin. 2nd Edition. People’s Literature Publishing House, 2012.
  6. Simulating opinion dynamics with networks of llm-based agents. arXiv preprint arXiv:2311.09618, 2023.
  7. Efficient and effective text encoding for chinese llama and alpaca. arXiv preprint arXiv:2304.08177, 2023. URL https://arxiv.org/abs/2304.08177.
  8. Enhancing chat language models by scaling high-quality instructional conversations. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, pages 3029–3051, 2023.
  9. Iason Gabriel. Artificial intelligence, values, and alignment. Minds and machines, 30(3):411–437, 2020.
  10. Ai and the transformation of social science research. Science, 380(6650):1108–1109, 2023.
  11. S Güver and Renate Motschnig. Effects of diversity in teams and workgroups: A qulitative systematic review. International Journal of Business, Humanities and Technology, 2017.
  12. Pre-trained models: Past, present and future. AI Open, 2:225–250, 2021.
  13. Reference-free monolithic preference optimization with odds ratio. arXiv preprint arXiv:2403.07691, 2024.
  14. Minicpm: Unveiling the potential of small language models with scalable training strategies. arXiv preprint arXiv:2404.06395, 2024.
  15. Personallm: Investigating the ability of large language models to express personality traits. In Findings of the Association for Computational Linguistics: NAACL 2024, pages 3605–3627, 2024.
  16. The benefits, risks and bounds of personalizing the alignment of large language models to individuals. Nature Machine Intelligence, pages 1–10, 2024.
  17. Chin-Yew Lin. Rouge: A package for automatic evaluation of summaries. In Text summarization branches out, pages 74–81, 2004.
  18. Trustworthy llms: a survey and guideline for evaluating large language models’ alignment. arXiv preprint arXiv:2308.05374, 2023.
  19. Gardner Murphy. Personality: A biosocial approach to origins and structure. Harper & Brothers, 1947.
  20. Sylvia Nasar. A beautiful mind. Simon and Schuster, 2011.
  21. Large dual encoders are generalizable retrievers. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pages 9844–9855, 2022.
  22. Training language models to follow instructions with human feedback. Advances in neural information processing systems, 35:27730–27744, 2022.
  23. Bleu: a method for automatic evaluation of machine translation. In Proceedings of the 40th annual meeting of the Association for Computational Linguistics, pages 311–318, 2002.
  24. Influencing human–ai interaction by priming beliefs about ai can increase perceived trustworthiness, empathy and effectiveness. Nature Machine Intelligence, 5(10):1076–1086, 2023.
  25. What makes an ai device human-like? the role of interaction quality, empathy and perceived psychological anthropomorphic characteristics in the acceptance of artificial intelligence in the service industry. Computers in Human Behavior, 122:106855, 2021.
  26. Language models are unsupervised multitask learners. OpenAI blog, 1(8):9, 2019.
  27. Direct preference optimization: Your language model is secretly a reward model. Advances in Neural Information Processing Systems, 36, 2024.
  28. Proximal policy optimization algorithms. arXiv preprint arXiv:1707.06347, 2017.
  29. Role play with large language models. Nature, 623(7987):493–498, 2023.
  30. Character-llm: A trainable agent for role-playing. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, pages 13153–13187, 2023.
  31. Llama: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971, 2023.
  32. A survey on large language model based autonomous agents. Frontiers of Computer Science, 18(6):1–26, 2024.
  33. Self-instruct: Aligning language models with self-generated instructions. In The 61st Annual Meeting Of The Association For Computational Linguistics, 2023a.
  34. Aligning large language models with human: A survey. arXiv preprint arXiv:2307.12966, 2023b.
  35. Rolellm: Benchmarking, eliciting, and enhancing role-playing abilities of large language models. arXiv preprint arXiv:2310.00746, 2023c.
  36. Emergent abilities of large language models. arXiv preprint arXiv:2206.07682, 2022.
  37. Simple synthetic data reduces sycophancy in large language models. arXiv e-prints, pages arXiv–2308, 2023.
  38. Diyi Yang. Human-ai interaction in the age of large language models. In Proceedings of the AAAI Symposium Series, volume 3, pages 66–67, 2024.
  39. Exploring the impact of instruction data scaling on large language models: An empirical study on real-world use cases. arXiv preprint arXiv:2303.14742, 2023.
  40. Knowledgeable preference alignment for llms in domain-specific question answering. arXiv preprint arXiv:2311.06503, 2023.
  41. A survey of large language models. arXiv preprint arXiv:2303.18223, 2023.
  42. Characterglm: Customizing chinese conversational ai characters with large language models. arXiv preprint arXiv:2311.16832, 2023a.
  43. Sotopia: Interactive evaluation for social intelligence in language agents. In The 12th International Conference on Learning Representations, 2023b.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Zheni Zeng (15 papers)
  2. Jiayi Chen (63 papers)
  3. Huimin Chen (15 papers)
  4. Yukun Yan (39 papers)
  5. Yuxuan Chen (80 papers)
  6. Zhiyuan Liu (433 papers)
  7. Maosong Sun (337 papers)
  8. Zhenghao Liu (77 papers)
Citations (2)
Youtube Logo Streamline Icon: https://streamlinehq.com