Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Long-Tail Crisis in Nearest Neighbor Language Models (2503.22426v1)

Published 28 Mar 2025 in cs.CL

Abstract: The $k$-nearest-neighbor LLM ($k$NN-LM), one of the retrieval-augmented LLMs, improves the perplexity for given text by directly accessing a large datastore built from any text data during inference. A widely held hypothesis for the success of $k$NN-LM is that its explicit memory, i.e., the datastore, enhances predictions for long-tail phenomena. However, prior works have primarily shown its ability to retrieve long-tail contexts, leaving the model's performance remain underexplored in estimating the probabilities of long-tail target tokens during inference. In this paper, we investigate the behavior of $k$NN-LM on low-frequency tokens, examining prediction probability, retrieval accuracy, token distribution in the datastore, and approximation error of the product quantization. Our experimental results reveal that $k$NN-LM does not improve prediction performance for low-frequency tokens but mainly benefits high-frequency tokens regardless of long-tail contexts in the datastore.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Yuto Nishida (5 papers)
  2. Makoto Morishita (20 papers)
  3. Hiroyuki Deguchi (8 papers)
  4. Hidetaka Kamigaito (62 papers)
  5. Taro Watanabe (76 papers)