Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
162 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Examination and Extension of Strategies for Improving Personalized Language Modeling via Interpolation (2006.05469v1)

Published 9 Jun 2020 in cs.CL and cs.LG

Abstract: In this paper, we detail novel strategies for interpolating personalized LLMs and methods to handle out-of-vocabulary (OOV) tokens to improve personalized LLMs. Using publicly available data from Reddit, we demonstrate improvements in offline metrics at the user level by interpolating a global LSTM-based authoring model with a user-personalized n-gram model. By optimizing this approach with a back-off to uniform OOV penalty and the interpolation coefficient, we observe that over 80% of users receive a lift in perplexity, with an average of 5.2% in perplexity lift per user. In doing this research we extend previous work in building NLIs and improve the robustness of metrics for downstream tasks.

Citations (4)

Summary

We haven't generated a summary for this paper yet.