Papers
Topics
Authors
Recent
2000 character limit reached

Persona-Driven Patient Dialogue

Updated 27 December 2025
  • The paper introduces personalized dialogue systems that integrate patient-specific profiles into multi-turn interactions to boost trust, factuality, and training realism.
  • Methodologies include rule-based templates, neural prompt-tuning, and memory-augmented refinement to ensure stylistic consistency and contextual relevance.
  • Findings demonstrate that dynamic persona alignment improves realism and consistency in simulated patient consultations for both clinical training and research.

Persona-driven patient dialogue is a class of conversational agent methodologies and systems in which patient-specific attributes—ranging from demographic or medical profiles to nuanced personality, linguistic capabilities, and emotional states—are explicitly modeled and injected into multi-turn medical dialogues. This paradigm enables more contextually relevant, trust-building, and pedagogically authentic consultations, supporting both clinical communication training and virtual patient simulation at scale. Approaches span rule-based persona stylization, controllable neural generation, prompt-tuning, uncertainty-aware preference elicitation, memory-augmented refinement, and large-scale persona simulation frameworks, with evaluation covering fidelity to profile, emotional expressivity, medical factuality, and human-likeness.

1. Foundational Concepts in Persona-Driven Patient Dialogue

Persona-driven dialogue integrates explicit or latent user models into response generation, modulating content, tone, and information complexity in response to medical, psychological, and social characteristics. Vered et al. (2019) formalize four core modules: (1) domain knowledge extraction for valid medical explanations, (2) patient profile inference encompassing demographic, emotional, cognitive, and medical history vectors Up={hp,kp,ep,...}U_p = \{h_p, k_p, e_p, ...\}, (3) personalized explanation generation via mappings Ep=f(K,Up)E_p = f(K, U_p) from all candidate knowledge KK to patient-relevant “subsets,” and (4) an adaptive user interface controlling modality and pacing (Vered et al., 2019). These elements define the functional requirements for any system that dynamically grounds communication in patient-specific context.

Persona definitions vary from hand-crafted template sets (profile, name, stylistic markers (Hwang et al., 2021)) to learned key-value embeddings, sparse attribute sequences, topic-augmented vocabularies, and soft prefix vectors (e.g., PRk×dP \in \mathbb{R}^{k \times d} in prompt-tuning (Kasahara et al., 2022)). Persona can encode static facts (gender, age, diagnosis), mutable psychological signals (anxiety, optimism level), fidelity of recall/memory, and idiosyncratic communication profiles (verbosity, skepticism, confusion phase).

Significance in clinical settings includes enhanced trust, improved adherence, and realistic skill acquisition for trainees; in NLP, it enables statistical control and evaluation of personalization, coherence, and style consistency over multi-turn, multi-session dialogues.

2. Approaches to Persona Representation and Injection

Static and Template-Based Persona Modeling

Early applied work implements persona by hand-crafting dialogue templates stylized according to real-world interlocutors (family member, doctor), capturing micro-markers (endearments, hedges, emojis, intentional typos) and anchoring to profile fields and conversational logs (Hwang et al., 2021). These systems eschew vector embeddings and learnable modules, instead maintaining persona fidelity through manual authoring and stylistic slot-filling. This approach demonstrated statistically significant effects on perceived intimacy and trust (e.g., Family persona raised intimacy, Doctor persona increased trust; F(2,30)=19.42, p<0.001 for intimacy).

Neural Persona-Conditioned Generation

Neural dialogue systems encode personas as structured attribute sets or learned embeddings. A canonical framework uses a Transformer encoder–decoder architecture initialized from large-scale LLMs, where the encoder ingests both dialogue context CC and target persona T={ki,vi}T = \{\langle k_i, v_i\rangle\}, mapping entity attributes to distinct memory slots (Zheng et al., 2019). Persona features are merged during decoding using parallel multi-head attention routing:

Omerge=αOT+(1α)OC+OprevO_\text{merge} = \alpha \cdot O_T + (1-\alpha) O_C + O_\text{prev}

with α[0,1]\alpha \in [0,1] predicted per-turn as the “persona weight,” dynamically balancing generic and persona-grounded content.

Alternate approaches exploit soft prompts PRk×dP \in \mathbb{R}^{k\times d} prepended to all inputs as learned prefix vectors; only PP is updated during training, while LM weights remain fixed. Persona sentences (demographics, medical conditions, needs) are directly encoded into these soft vectors (Kasahara et al., 2022).

Neural topical expansion frameworks (PEE) leverage a VAE-based topic model for persona exploration, expanding sparse persona descriptions with strongly correlated medical vocabulary before multi-source fusion in a hierarchical encoder–decoder pipeline (Xu et al., 2020). Persona-alignment and consistency are supervised using auxiliary losses (P-Match, P-BoWs).

Explicit Persona Axes and Controlled Diversity

PatientSim operationalizes multidimensional persona descriptors across four axes: personality (e.g., neutral, anxious), language proficiency (CEFR A/B/C), medical recall, and confusion level, combinatorially yielding 37 discrete patient profiles (Kyung et al., 23 May 2025). Profiles are constructed from real-world MIMIC-ED/IV data and narrative fields, then injected as persona descriptors in every prompt to systematically control linguistic, cognitive, and behavioral response variance.

3. Memory-Augmented and Two-Stage Persona Alignment

Multi-session persona preservation and controllable response naturalness are addressed by memory-augmented post-aligning systems. The Post Persona Alignment (PPA) paradigm (Chen et al., 13 Jun 2025) decouples response generation from persona grounding: it generates a draft response RgR_g using only prior context, then retrieves top-kk persona memory triples from a persona/history memory store MM by cosine similarity:

s(Rg,mi)=cosine(Embed(Rg),Embed(mi))s(R_g, m_i) = \text{cosine}(Embed(R_g), Embed(m_i))

Persona-aligned refinement is performed in a second LLM pass, integrating RgR_g, MkM_k, and context to ensure injection of patient-specific facts, emotional cues, and stylistic consistency. Empirical results show doubled consistency (C-Score up to 0.456), increased diversity (entropy), and improved persona-F1 versus standard retrieve-then-generate pipelines.

This architecture supports session-spanning consistency (e.g., management of allergies or chronic diagnoses) and fine control over the “persona signal” via triples and retrieval thresholds.

4. Uncertainty, Knowledge Gap, and Dynamic Persona Elicitation

The CPER framework (Baskar et al., 16 Mar 2025) addresses the “persona knowledge gap,” i.e., the mismatch between model beliefs about the patient’s persona and uncertainty about required context for personalized support. At each turn tt, CPER quantifies model confidence (uncertainty utu_t) and alignment with persona history (WCMI), then computes:

KGt=1+αutβWCMI(pt,Pattended)KG_t = 1 + \alpha u_t - \beta\mathrm{WCMI}(p_t,P_{\mathrm{attended}})

If the gap exceeds a threshold, targeted clarification prompts (ftf_t) are generated and updated persona vectors are re-integrated. This approach leads to large human- and LLM-preference gains (+27% human preference on mental health support ESConv dataset), particularly over long (12+ turn) dialogues, by explicitly querying and closing critical knowledge gaps via embedding-based information fusion and multi-turn refinement.

5. Prompt Engineering and Evaluation in LLM Persona Simulation

LLMs have demonstrated effective persona “role-playing” via prompt engineering alone, obviating retraining or finetuning for most accuracy-focused simulations (Ma et al., 12 Jan 2025, Kasahara et al., 2022, Bodonhelyi et al., 28 Mar 2025). Key design elements include:

  • Isolated persona priming: single-attribute prompts (e.g., education level) maximize alignment to human reference responses (accuracy up to 88% for high-education prompts).
  • In-context instructions and output constraints: instructing models to output only labels/short answers reduces variance and hallucination. Over-customization with multi-attribute prompts degrades fidelity.
  • Behavioral scripting: JSON-structured system and author’s notes encode role, mood, adversity, communication style, and language sophistication (Bodonhelyi et al., 28 Mar 2025).

Evaluation is multi-faceted, with measures including factual entailment (per-sentence and profile extraction), persona-fidelity (Likert scale, clinician rating), expressivity (emotion and sentiment analysis), dialogue diversity, and automated consistency metrics (C-Score, Persona-F1). PatientSim’s Llama 3.3 70B, for example, achieves $3.89/4$ mean persona fidelity across axes, 79.6%79.6\% entailment, and high unsupported-information plausibility with inter-rater reliability AC1>0.86AC_1 > 0.86 (Kyung et al., 23 May 2025).

6. Application to Medical Training and Realistic Virtual Patients

Persona-driven dialogue is central to the design of virtual simulated patients (VSPs) and communication training tools (Marez et al., 20 Dec 2025). Key components include:

  • Persona initialization: Big-Five personality vectors B[0,5]5B \in [0,5]^5 jointly determine dialogue prompts PtextP_\text{text}.
  • Adaptive persona state updates: At each turn tt, persona state is updated as pt=αpt1+βφ(xt,ht1)p_t = \alpha p_{t-1} + \beta \varphi(x_t, h_{t-1}), reflecting observed utterance content.
  • Controlled natural language generation: Multi-module pipelines integrate conditional retrieval (RAG), explicit persona/role post-processing, and scenario-linked vignettes for consistent, medically accurate, and difficulty-calibrated training interactions.
  • Standards-based and example-rich feedback: After each session, detailed assessment vectors (e.g., MIRS scale, clinical reasoning categories) with traceable dialogue excerpts are automatically produced to anchor student feedback.

In addition, advanced prompt engineering enables simulation of challenging archetypes (accuser, rationalizer) across languages, emotion spectra, and patient narratives with authentic behavioral cues, equipping trainees to navigate difficult real-world scenarios (Bodonhelyi et al., 28 Mar 2025).

Best practices include periodic persona-reminder prompts, dynamic context reinforcement, explicit behavioral mechanisms (e.g., stubbornness, hesitancy), and session-wise calibration to maintain high-fidelity persona signals.

7. Limitations, Open Challenges, and Best Practices

Open challenges in persona-driven patient dialogue include:

  • Theory-of-mind scalability: Efficient, real-time updating of multidimensional patient states UpU_p amid limited annotated data and privacy constraints (Vered et al., 2019).
  • Knowledge gap minimization: Dynamic uncertainty estimation and clarification loop to ensure alignment between model persona estimates and patient reality (Baskar et al., 16 Mar 2025).
  • Equity and bias: LLM-based personas reflect training-data biases and underperform for underrepresented attributes (e.g., low/female education profiles) (Ma et al., 12 Jan 2025).
  • Long-term consistency: Management of cumulative memory/facts for multi-session, multi-user consultation (Chen et al., 13 Jun 2025).
  • Clinical safety and validation: Hallucination minimization, factual grounding, and safe-harbor fallback, especially in high-consequence domains (Kasahara et al., 2022, Kyung et al., 23 May 2025).
  • Multimodality and human-in-the-loop oversight: Extension to non-textual signals (prosody, expressions) and iterative refinement with real human feedback.

Recommended best practices, as distilled across studies, include single-attribute persona prompts for high alignment, separate evaluation of factual and perceptual tasks, controlled output templates, uncertainty-handling instructions, regular re-validation with human data, and modular separation of scenario, interaction, and assessment control (Ma et al., 12 Jan 2025, Marez et al., 20 Dec 2025, Chen et al., 13 Jun 2025).


Key references:

These works collectively define, operationalize, and empirically validate the design space and future directions of persona-driven patient dialogue.

Whiteboard

Topic to Video (Beta)

Follow Topic

Get notified by email when new papers are published related to Persona-Driven Patient Dialogue.