Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Pearl: Personalizing Large Language Model Writing Assistants with Generation-Calibrated Retrievers (2311.09180v2)

Published 15 Nov 2023 in cs.CL, cs.HC, and cs.IR

Abstract: Powerful LLMs have facilitated the development of writing assistants that promise to significantly improve the quality and efficiency of composition and communication. However, a barrier to effective assistance is the lack of personalization in LLM outputs to the author's communication style, specialized knowledge, and values. In this paper, we address this challenge by proposing Pearl, a LLM writing assistant personalized with a retriever that is trained to be generation-calibrated for personalization. Generation calibration ensures that our retriever selects historic user authored documents to augment an LLM prompt such that they are likely to help an LLM generation better adhere to a users' preferences. We propose two key novelties for training such a retriever: (1) A training data selection method that identifies user requests likely to benefit from personalization and documents that provide that benefit; and (2) A scale-calibrating KL-divergence objective that ensures that our retriever scores remain proportional to the downstream generation quality from using the document for personalized generation. In a series of holistic evaluations, we demonstrate the effectiveness of Pearl in generating long-form texts on multiple social media datasets. Finally, we demonstrate how a generation-calibrated retriever can double as a performance predictor -- detecting low quality retrieval, and improving potentially under-performing outputs via revision with LLMs.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (10)
  1. Sheshera Mysore (15 papers)
  2. Zhuoran Lu (7 papers)
  3. Mengting Wan (24 papers)
  4. Longqi Yang (28 papers)
  5. Steve Menezes (1 paper)
  6. Tina Baghaee (1 paper)
  7. Emmanuel Barajas Gonzalez (1 paper)
  8. Jennifer Neville (57 papers)
  9. Tara Safavi (16 papers)
  10. Bahareh Sarrafzadeh (8 papers)
Citations (22)
X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets