Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash 99 tok/s
Gemini 2.5 Pro 60 tok/s Pro
GPT-5 Medium 32 tok/s
GPT-5 High 27 tok/s Pro
GPT-4o 102 tok/s
GPT OSS 120B 461 tok/s Pro
Kimi K2 227 tok/s Pro
2000 character limit reached

PRIME: Large Language Model Personalization with Cognitive Memory and Thought Processes (2507.04607v2)

Published 7 Jul 2025 in cs.CL and cs.AI

Abstract: LLM personalization aims to align model outputs with individuals' unique preferences and opinions. While recent efforts have implemented various personalization methods, a unified theoretical framework that can systematically understand the drivers of effective personalization is still lacking. In this work, we integrate the well-established cognitive dual-memory model into LLM personalization, by mirroring episodic memory to historical user engagements and semantic memory to long-term, evolving user beliefs. Specifically, we systematically investigate memory instantiations and introduce a unified framework, PRIME, using episodic and semantic memory mechanisms. We further augment PRIME with a novel personalized thinking capability inspired by the slow thinking strategy. Moreover, recognizing the absence of suitable benchmarks, we introduce a dataset using Change My View (CMV) from Reddit, specifically designed to evaluate long-context personalization. Extensive experiments validate PRIME's effectiveness across both long- and short-context scenarios. Further analysis confirms that PRIME effectively captures dynamic personalization beyond mere popularity biases.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Summary

  • The paper introduces PRIME, a framework that integrates episodic and semantic memory to personalize LLM outputs based on user history and evolving beliefs.
  • It employs personalized thinking and chain-of-thought generation using cognitive memory models, achieving significant response relevance in experiments on the CMV dataset.
  • Experimental results confirm that authentic user profiles are crucial, as performance sharply drops with mismatched histories, underscoring PRIME’s user-centric design.

PRIME: LLM Personalization with Cognitive Memory and Thought Processes

Introduction to PRIME

The paper "PRIME: LLM Personalization with Cognitive Memory and Thought Processes" introduces a novel framework for personalizing LLMs. This framework, dubbed PRIME, integrates dual-memory cognitive models comprising episodic and semantic memories, thereby enabling LLMs to better align outputs with users' unique preferences and opinions. PRIME goes beyond existing personalization methods by offering a unified theoretical framework that combines historical user engagements and long-term evolving beliefs.

Framework Components

PRIME leverages a dual-memory model to systematically capture and utilize user-specific historical and semantic information. The framework is inspired by well-established cognitive theories that differentiate between episodic memory—specific personal experiences—and semantic memory—abstract knowledge and beliefs. Figure 1

Figure 1: Overview of our unified framework, PRIME, inspired by dual-memory model.

  1. Episodic Memory (EM): Stores interaction-specific data, capturing recent user engagements. The memory reading mechanism employs strategies like recent and relevant history recalls.
  2. Semantic Memory (SM): Encodes generalized user preferences abstracted from historical data. This can be instantiated via parametric training approaches such as fine-tuning and input-only training, with semantic abstractions realized through model parameters.
  3. Integration and Personalization: The paper introduces personalized thinking, an augmented capability of PRIME drawn from slow thinking strategies, allowing for generation of chain-of-thoughts that reflect personalized reasoning paths.

Benchmarking and Dataset

Recognizing the lack of suitable benchmarks for evaluating long-context personalization, the authors introduce a dataset derived from the Change My View (CMV) Reddit forum. This dataset is specifically tailored to test the long- and short-context personalization capabilities of LLMs. CMV data enables a ranking-based recommendation task, where the objective is identifying responses that effectively alter users' points of view.

Experimental Results

Empirical studies on CMV data and existing benchmarks demonstrate that:

  1. Semantic Memory Dominance: Semantic memory usually provides a more robust personalization compared to episodic memory alone.
  2. PRIME Effectiveness: By integrating both memory types with personalized thinking, PRIME consistently delivers responses that are both contextually relevant and aligned with user-specific long-term beliefs.
  3. Profile Replacement Sensitivity: Experiments confirm PRIME's sensitivity to authentic user history, as performance notably dips when personalized memory is replaced with other users' history. This underscores PRIME's dynamic personalization rather than reliance on bandwagon biases. Figure 2

    Figure 2: Average performance under five user-profile replacement conditions. Performance drops sharply when a target user's profile is replaced, confirming the faithfulness of PRIME to user history.

Scalability and Implementation

PRIME's scalable nature is highlighted by its compatibility with different model families and sizes. The framework leverages efficient methods like LoRA for memory instantiation, ensuring flexibility in user adaptation, even with limited historical data — often termed the "cold-start" challenge. Figure 3

Figure 3: Average performance for Train-free Personalized Thinking paper. Think refers to our train-free thinking approach.

Conclusions and Future Work

The introduction of PRIME as a unified framework for LLM personalization marks a significant advancement in the field, integrating cognitive memory models to achieve more nuanced user alignment. Future work will likely focus on expanding PRIME's applications across broader domains and enhancing its capabilities through further exploration of memory integration and personalized reasoning techniques. The adaptability and robustness of PRIME suggest promising directions for ongoing research in scalable, user-centric AI systems.

Ai Generate Text Spark Streamline Icon: https://streamlinehq.com

Paper Prompts

Sign up for free to create and run prompts on this paper using GPT-5.

Don't miss out on important new AI/ML research

See which papers are being discussed right now on X, Reddit, and more:

“Emergent Mind helps me see which AI papers have caught fire online.”

Philip

Philip

Creator, AI Explained on YouTube