Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
60 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Integrating Summarization and Retrieval for Enhanced Personalization via Large Language Models (2310.20081v1)

Published 30 Oct 2023 in cs.CL, cs.AI, and cs.IR

Abstract: Personalization, the ability to tailor a system to individual users, is an essential factor in user experience with NLP systems. With the emergence of LLMs, a key question is how to leverage these models to better personalize user experiences. To personalize a LLM's output, a straightforward approach is to incorporate past user data into the LLM prompt, but this approach can result in lengthy inputs exceeding limitations on input length and incurring latency and cost issues. Existing approaches tackle such challenges by selectively extracting relevant user data (i.e. selective retrieval) to construct a prompt for downstream tasks. However, retrieval-based methods are limited by potential information loss, lack of more profound user understanding, and cold-start challenges. To overcome these limitations, we propose a novel summary-augmented approach by extending retrieval-augmented personalization with task-aware user summaries generated by LLMs. The summaries can be generated and stored offline, enabling real-world systems with runtime constraints like voice assistants to leverage the power of LLMs. Experiments show our method with 75% less of retrieved user data is on-par or outperforms retrieval augmentation on most tasks in the LaMP personalization benchmark. We demonstrate that offline summarization via LLMs and runtime retrieval enables better performance for personalization on a range of tasks under practical constraints.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Chris Richardson (2 papers)
  2. Yao Zhang (537 papers)
  3. Kellen Gillespie (3 papers)
  4. Sudipta Kar (19 papers)
  5. Arshdeep Singh (19 papers)
  6. Zeynab Raeesy (6 papers)
  7. Omar Zia Khan (3 papers)
  8. Abhinav Sethy (14 papers)
Citations (4)