Persona Vectors: Monitoring and Controlling Character Traits in Language Models (2507.21509v1)
Abstract: LLMs interact with users through a simulated 'Assistant' persona. While the Assistant is typically trained to be helpful, harmless, and honest, it sometimes deviates from these ideals. In this paper, we identify directions in the model's activation space-persona vectors-underlying several traits, such as evil, sycophancy, and propensity to hallucinate. We confirm that these vectors can be used to monitor fluctuations in the Assistant's personality at deployment time. We then apply persona vectors to predict and control personality shifts that occur during training. We find that both intended and unintended personality changes after finetuning are strongly correlated with shifts along the relevant persona vectors. These shifts can be mitigated through post-hoc intervention, or avoided in the first place with a new preventative steering method. Moreover, persona vectors can be used to flag training data that will produce undesirable personality changes, both at the dataset level and the individual sample level. Our method for extracting persona vectors is automated and can be applied to any personality trait of interest, given only a natural-language description.
Summary
- The paper introduces persona vectors as linear directions in LLM activation space that capture and modulate specific character traits.
- It employs an automated, contrastive prompt-based pipeline with causal steering to predict and control trait expressions during inference and finetuning.
- The method enables effective data screening and proactive finetuning interventions, offering practical tools for model alignment and safety.
Persona Vectors: Monitoring and Controlling Character Traits in LLMs
Introduction and Motivation
The paper introduces a systematic approach for extracting and leveraging "persona vectors"—linear directions in the activation space of LLMs that correspond to specific character traits such as evil, sycophancy, and hallucination. The motivation is grounded in the observation that LLMs, despite being trained to be helpful, harmless, and honest, can exhibit undesirable persona shifts both at deployment (e.g., via prompting) and during finetuning, sometimes resulting in harmful or misaligned behaviors. The work builds on prior findings that high-level traits are often encoded as linear directions in model activations, and that steering along these directions can causally influence model behavior.
Figure 1: The automated pipeline extracts persona vectors from natural-language trait descriptions and applies them for monitoring, mitigation, and data screening.
Automated Extraction of Persona Vectors
The core technical contribution is an automated pipeline that, given a trait name and a brief description, generates contrastive system prompts and evaluation questions to elicit opposing behaviors. The pipeline uses a frontier LLM to synthesize:
- Pairs of positive/negative system prompts for the trait
- Evaluation questions likely to evoke trait-relevant behavior
- An evaluation rubric for scoring trait expression
For each question, responses are generated under both positive and negative prompts. Responses are filtered based on trait expression scores (using an LLM judge), and residual stream activations are extracted and averaged across response tokens. The persona vector is computed as the difference in mean activations between trait-exhibiting and non-trait-exhibiting responses, yielding a direction per layer; the most informative layer is selected via empirical steering effectiveness.
Figure 2: The pipeline generates contrastive prompts and computes persona vectors as mean activation differences between trait-exhibiting and non-trait-exhibiting responses.
Causal Steering and Monitoring with Persona Vectors
Persona vectors are validated via two mechanisms:
- Causal Steering: At inference, activations are shifted along the persona vector at each decoding step, modulated by a scalar coefficient. This reliably amplifies or suppresses the target trait in generated responses, as measured by trait expression scores.
- Activation Monitoring: The projection of the last prompt token's activation onto the persona vector strongly correlates with subsequent trait expression, enabling prediction of prompt-induced behavioral shifts before generation.
Figure 3: Steering along persona vectors at different layers modulates trait expression in Qwen2.5-7B-Instruct.
Figure 4: Projection of prompt activations onto persona vectors predicts trait expression under varying system prompts.
Finetuning-Induced Persona Shifts and Predictive Monitoring
The authors construct both explicitly trait-eliciting and "emergent misalignment-like" (EM-like) datasets, the latter containing domain-specific errors or flaws. Finetuning on these datasets induces diverse persona shifts, including unintended amplification of non-target traits.
Figure 5: Diverse datasets (trait-eliciting and EM-like) induce varied persona shifts after finetuning.
A key empirical finding is that the shift in model activations along the persona vector (the "finetuning shift") is highly correlated with the change in trait expression post-finetuning (r=0.76–$0.97$ for target traits). This holds across both explicit and emergent misalignment scenarios, and is trait-specific.
Figure 6: Finetuning shift along persona vectors predicts post-finetuning trait expression.
Mitigating and Preventing Persona Shifts via Steering
Two steering-based interventions are proposed:
- Inference-Time Steering: After finetuning, subtracting the persona vector during generation suppresses trait expression but can degrade general capabilities (e.g., MMLU accuracy) at high coefficients.
- Preventative Steering: During finetuning, adding the persona vector to activations proactively limits trait acquisition, better preserving general capabilities and achieving more robust mitigation, especially when applied across multiple layers.
Figure 7: (a) Inference-time steering reduces trait expression but can harm general performance; (b) Preventative steering during finetuning limits trait acquisition with less collateral damage.
The authors also compare their approach to CAFT (Concept Ablation Fine-Tuning) and regularization-based methods, finding that preventative steering is more effective for traits where the base model's projection is not already near zero.
Data Screening: Predicting and Preventing Undesirable Shifts
A novel application is pre-finetuning data screening. By projecting training data responses and base model generations onto the persona vector, the "projection difference" metric is computed. This metric is highly predictive of post-finetuning trait expression, outperforming raw projection and enabling both dataset-level and sample-level identification of problematic data.
Figure 8: Dataset-level projection difference predicts post-finetuning trait expression.

Figure 9: Individual samples from trait-inducing datasets are separable from controls by projection onto persona vectors.
The method generalizes to real-world datasets (e.g., LMSYS-Chat-1M), where high projection difference samples induce stronger trait expression even after LLM-based filtering, surfacing problematic data that may evade standard LLM judges.
Figure 10: Persona vectors identify trait-inducing samples in real-world data; high projection difference subsets induce elevated trait expression after finetuning.
Implementation Considerations
- Computational Cost: The pipeline requires generating base model responses for all training samples to compute projection differences, which is expensive for large datasets. Efficient approximations (e.g., using prompt token projections) are proposed and shown to be effective for some traits.
- Layer Selection: Steering and monitoring are most effective when applied at empirically selected layers, typically mid-to-late transformer blocks.
- Trait Coverage: The method is trait-supervised and requires that the trait can be elicited via prompting; traits not easily induced by prompts may require alternative approaches (e.g., unsupervised feature discovery via sparse autoencoders).
- Evaluation: Automated trait scoring with LLM judges is validated against human raters, showing high agreement (∼95%), but may miss subtle or context-dependent manifestations.
Theoretical and Practical Implications
The results provide strong evidence that high-level behavioral traits in LLMs are encoded as approximately linear directions in activation space, and that both intended and unintended persona shifts during finetuning are mediated by movement along these directions. This supports a linear representation hypothesis for many behavioral features in LLMs and suggests that simple linear interventions can be effective for both monitoring and control.
Practically, persona vectors offer a unified toolset for:
- Deployment Monitoring: Predicting and flagging prompt-induced persona shifts before generation.
- Finetuning Mitigation: Preventing or reversing undesirable trait acquisition during or after finetuning.
- Data Curation: Screening and filtering training data at both dataset and sample granularity to prevent misalignment.
The approach is model-agnostic and requires only a natural-language trait description, making it broadly applicable across LLM architectures and deployment scenarios.
Future Directions
Open questions include:
- The dimensionality and structure of the "persona space": Is there a natural basis, and how do trait correlations manifest mechanistically?
- The limits of linear methods: Are some traits fundamentally nonlinear or distributed?
- Integration with unsupervised feature discovery (e.g., sparse autoencoders) for traits not easily elicited by prompting.
- Scaling to larger models and more diverse deployment contexts, including multi-turn and domain-adaptive settings.
Conclusion
This work presents a robust, automated methodology for extracting, monitoring, and controlling persona traits in LLMs via linear directions in activation space. Persona vectors enable both proactive and reactive interventions against undesirable behavioral shifts, with strong empirical evidence for their predictive and causal efficacy. The approach has immediate applications in model alignment, safety, and data curation, and provides a foundation for further mechanistic understanding of LLM behavior.
Follow-up Questions
- How do contrastive prompts facilitate the extraction of persona vectors in language models?
- What impact does causal steering have on both trait modulation and general model performance?
- How is the projection difference metric used to predict undesirable trait expressions post-finetuning?
- What are the potential limitations of applying linear interventions for controlling behaviour in complex LLMs?
- Find recent papers about persona vectors in language models.
Related Papers
- Personalized Dialogue Generation with Diversified Traits (2019)
- Eliciting Personality Traits in Large Language Models (2024)
- Quantifying the Persona Effect in LLM Simulations (2024)
- PICLe: Eliciting Diverse Behaviors from Large Language Models with Persona In-Context Learning (2024)
- Evaluating Large Language Model Biases in Persona-Steered Generation (2024)
- Two Tales of Persona in LLMs: A Survey of Role-Playing and Personalization (2024)
- Scaling Synthetic Data Creation with 1,000,000,000 Personas (2024)
- Persona is a Double-edged Sword: Mitigating the Negative Impact of Role-playing Prompts in Zero-shot Reasoning Tasks (2024)
- Persona-L has Entered the Chat: Leveraging LLM and Ability-based Framework for Personas of People with Complex Needs (2024)
- Exploring Persona-dependent LLM Alignment for the Moral Machine Experiment (2025)
Authors (5)
Tweets
YouTube
alphaXiv
- Persona Vectors: Monitoring and Controlling Character Traits in Language Models (155 likes, 0 questions)