Papers
Topics
Authors
Recent
Detailed Answer
Quick Answer
Concise responses based on abstracts only
Detailed Answer
Well-researched responses based on abstracts and relevant paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses
Gemini 2.5 Flash
Gemini 2.5 Flash 84 tok/s
Gemini 2.5 Pro 45 tok/s Pro
GPT-5 Medium 28 tok/s Pro
GPT-5 High 21 tok/s Pro
GPT-4o 92 tok/s Pro
GPT OSS 120B 425 tok/s Pro
Kimi K2 157 tok/s Pro
2000 character limit reached

PsyMem: Fine-grained psychological alignment and Explicit Memory Control for Advanced Role-Playing LLMs (2505.12814v1)

Published 19 May 2025 in cs.CL and cs.AI

Abstract: Existing LLM-based role-playing methods often rely on superficial textual descriptions or simplistic metrics, inadequately modeling both intrinsic and extrinsic character dimensions. Additionally, they typically simulate character memory with implicit model knowledge or basic retrieval augment generation without explicit memory alignment, compromising memory consistency. The two issues weaken reliability of role-playing LLMs in several applications, such as trustworthy social simulation. To address these limitations, we propose PsyMem, a novel framework integrating fine-grained psychological attributes and explicit memory control for role-playing. PsyMem supplements textual descriptions with 26 psychological indicators to detailed model character. Additionally, PsyMem implements memory alignment training, explicitly trains the model to align character's response with memory, thereby enabling dynamic memory-controlled responding during inference. By training Qwen2.5-7B-Instruct on our specially designed dataset (including 5,414 characters and 38,962 dialogues extracted from novels), the resulting model, termed as PsyMem-Qwen, outperforms baseline models in role-playing, achieving the best performance in human-likeness and character fidelity.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Summary

  • The paper presents the PsyMem framework that enhances LLM role-playing by incorporating 26 psychological indicators and explicit memory control.
  • It employs a dual-stage training strategy on a curated dataset of 5,414 characters and 38,962 dialogues to improve character fidelity and human-likeness.
  • Experimental results demonstrate PsyMem’s superiority over baseline models in delivering dynamic, memory-driven role-playing interactions.

PsyMem: Fine-grained Psychological Alignment and Explicit Memory Control for Advanced Role-Playing LLMs

Introduction

The "PsyMem" framework addresses fundamental limitations in existing LLM-based role-playing systems, including superficial characterization and implicit memory modeling, which compromise reliability in applications like social simulations. By introducing detailed psychological attributes and explicit memory control, PsyMem significantly improves character fidelity and human-likeness in role-playing models. PsyMem enriches character modeling with 26 psychological indicators and aligns character responses with dynamically retrieved memory. Training on a specially curated dataset of 5,414 characters and 38,962 dialogues extracted from novels, the PsyMem-Qwen model demonstrates superior role-playing performance compared to baseline models. Figure 1

Figure 1: The performance comparison is conducted on two subsets: "Ordinary," consisting of 20 randomly selected characters from our test set, and "Famous," consisting of 20 well-known characters. The average attribute scores are calculated as the mean of the quantized scores across the 20 characters in each subset.

Dataset Architecture

The PsyMem framework introduces a comprehensive dataset architecture, integrating modern psychological attributes and explicit memory. The dataset $D_{\mathrm{RP}$ is organized around role profiles featuring 26 quantitative dimensions, dialogue contexts, queries, responses, and memory components. Data are extracted from 539 novels to maintain character authenticity and diversity. Figure 2

Figure 2: The genre distribution in the dataset.

Quantifiable Attributes and Memory Integration

PsyMem systematically categorizes character traits into latent psychological attributes and explicit behavioral patterns, leveraging frameworks like the Big Five Personality Model and Schwartz's Theory of Basic Values. Memory integration involves constructing character-specific knowledge graphs to enhance role fidelity, simulating memory dynamics in inference.

Training and Evaluation

The training process embodies a dual-stage strategy, incorporating basic role-play capacity development and specialized memory-augmented fine-tuning. This approach preserves general language understanding while optimizing character-specific role-playing capability. Evaluation employs a structured approach with criteria centered on character fidelity and character-independent capabilities, using GPT-4o assessment for credibility. Figure 3

Figure 3: The two stages training of role-playing LLM; fine-grained character profile and contextual memory enhance role-playing precision.

Experimental Results

PsyMem-Qwen and PsyMem-LLama models display marked improvements, achieving increased fidelity and human-likeness scores relative to leading LLM benchmarks. Results indicate significant advancements in memory application, showcasing PsyMem's effectiveness in yielding character-consistent, memory-anchored dialogue interactions. Figure 4

Figure 4: The impact of data size on model performance using the Pure Dove (PD) dataset.

Conclusion

PsyMem establishes a robust methodology for enhancing role-playing LLMs, bridging intrinsic and extrinsic character dimensions with explicit memory control. By fostering superior character consistency and aligning role-play responses with psychological frameworks, PsyMem offers a promising avenue for realistic social simulation applications. Future research could further refine these models, exploring more sophisticated cognitive dimensions and applications in diverse interactive media environments.

Ai Generate Text Spark Streamline Icon: https://streamlinehq.com

Paper Prompts

Sign up for free to create and run prompts on this paper using GPT-5.