Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Enhancing Role-playing Systems through Aggressive Queries: Evaluation and Improvement (2402.10618v2)

Published 16 Feb 2024 in cs.CL

Abstract: The advent of LLMs has propelled dialogue generation into new realms, particularly in the field of role-playing systems (RPSs). While enhanced with ordinary role-relevant training dialogues, existing LLM-based RPSs still struggle to align with roles when handling intricate and trapped queries in boundary scenarios. In this paper, we design the Modular ORchestrated Trap-setting Interaction SystEm (MORTISE) to benchmark and improve the role-playing LLMs' performance. MORTISE can produce highly role-relevant aggressive queries through the collaborative effort of multiple LLM-based modules, and formulate corresponding responses to create an adversarial training dataset via a consistent response generator. We select 190 Chinese and English roles to construct aggressive queries to benchmark existing role-playing LLMs. Through comprehensive evaluation, we find that existing models exhibit a general deficiency in role alignment capabilities. We further select 180 of the roles to collect an adversarial training dataset (named RoleAD) and retain the other 10 roles for testing. Experiments on models improved by RoleAD indicate that our adversarial dataset ameliorates this deficiency, with the improvements demonstrating a degree of generalizability in ordinary scenarios.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (48)
  1. Qwen technical report. ArXiv, abs/2309.16609.
  2. A synthetic data generation framework for grounded dialogues. In Annual Meeting of the Association for Computational Linguistics.
  3. Murray R. Barrick and Michael K. Mount. 1991. The big five personality dimensions and job performance: A meta-analysis. Personnel Psychology, 44:1–26.
  4. Colossal-ai: A unified deep learning system for large-scale parallel training. arXiv preprint arXiv:2110.14883.
  5. Gregory J Boyle. 1995. Myers-briggs type indicator (mbti): some psychometric limitations. Australian Psychologist, 30(1):71–74.
  6. Language models are few-shot learners. ArXiv, abs/2005.14165.
  7. Large language models meet harry potter: A bilingual dataset for aligning dialogue agents with characters.
  8. Enhancing chat language models by scaling high-quality instructional conversations. In Conference on Empirical Methods in Natural Language Processing.
  9. Design and evaluation of intelligent agent prototypes for assistance with focus and productivity at work. Proceedings of the 25th International Conference on Intelligent User Interfaces.
  10. Targen: Targeted data generation with large language models. ArXiv, abs/2310.17876.
  11. Measuring massive multitask language understanding. ArXiv, abs/2009.03300.
  12. Towards better instruction following language models for chinese: Investigating the impact of training data and evaluation. ArXiv, abs/2304.07854.
  13. Mistral 7b. arXiv preprint arXiv:2310.06825.
  14. Personallm: Investigating the ability of large language models to express big five personality traits.
  15. Chatharuhi: Reviving anime character in reality via large language model. ArXiv, abs/2308.09597.
  16. Camel: Communicative agents for "mind" exploration of large scale language model society. ArXiv, abs/2303.17760.
  17. A diversity-promoting objective function for neural conversation models. ArXiv, abs/1510.03055.
  18. Synthetic data generation with large language models for text classification: Potential and limitations. ArXiv, abs/2310.07849.
  19. Large language models are superpositions of all characters: Attaining arbitrary role-play via self-alignment.
  20. To infinity and beyond: Show-1 and showrunner agents in multi-agent simulations.
  21. Editing personality for llms. ArXiv, abs/2310.02168.
  22. OpenAI. 2022. Chatgpt: Optimizing language models for dialogue. OpenAI Blog.
  23. OpenAI. 2023. Gpt-4 technical report.
  24. Dialogbench: Evaluating llms as human-like dialogue systems. ArXiv, abs/2311.01677.
  25. Keyu Pan and Yawen Zeng. 2023. Do llms possess a personality? making the mbti test an amazing evaluation for large language models. ArXiv, abs/2307.16180.
  26. Generative agents: Interactive simulacra of human behavior. Proceedings of the 36th Annual ACM Symposium on User Interface Software and Technology.
  27. Personality traits in large language models. ArXiv, abs/2307.00184.
  28. Lamp: When large language models meet personalization. ArXiv, abs/2304.11406.
  29. Role play with large language models. Nature, 623:493 – 498.
  30. Character-LLM: A trainable agent for role-playing. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, pages 13153–13187, Singapore. Association for Computational Linguistics.
  31. Roleeval: A bilingual role evaluation benchmark for large language models.
  32. A zero-shot approach for multi-user task-oriented dialog generation. In International Conference on Natural Language Generation.
  33. Llama: Open and efficient foundation language models. ArXiv, abs/2302.13971.
  34. Llama 2: Open foundation and fine-tuned chat models. ArXiv, abs/2307.09288.
  35. Charactereval: A chinese benchmark for role-playing conversational agent evaluation.
  36. Openchat: Advancing open-source language models with mixed-quality data. ArXiv, abs/2309.11235.
  37. Does role-playing chatbots capture the character personalities? assessing personality traits for role-playing chatbots. ArXiv, abs/2310.17976.
  38. Self-instruct: Aligning language models with self-generated instructions. In Annual Meeting of the Association for Computational Linguistics.
  39. Rolellm: Benchmarking, eliciting, and enhancing role-playing abilities of large language models. ArXiv, abs/2310.00746.
  40. Multi-party chat: Conversational agents in group settings with humans and models. ArXiv, abs/2304.13835.
  41. Baize: An open-source chat model with parameter-efficient tuning on self-chat data. ArXiv, abs/2304.01196.
  42. Cvalues: Measuring the values of chinese large language models from safety to responsibility. ArXiv, abs/2307.09705.
  43. Baichuan 2: Open large-scale language models. ArXiv, abs/2309.10305.
  44. Fuzzllm: A novel and universal fuzzing framework for proactively discovering jailbreak vulnerabilities in large language models. ArXiv, abs/2309.05274.
  45. Large language model as attributed training data generator: A tale of diversity and bias. ArXiv, abs/2306.15895.
  46. Bertscore: Evaluating text generation with bert. ArXiv, abs/1904.09675.
  47. Judging llm-as-a-judge with mt-bench and chatbot arena. ArXiv, abs/2306.05685.
  48. Characterglm: Customizing chinese conversational ai characters with large language models. ArXiv, abs/2311.16832.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Yihong Tang (24 papers)
  2. Jiao Ou (8 papers)
  3. Che Liu (59 papers)
  4. Fuzheng Zhang (60 papers)
  5. Di Zhang (230 papers)
  6. Kun Gai (125 papers)
Citations (4)
X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets