Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

$k$NN-Adapter: Efficient Domain Adaptation for Black-Box Language Models (2302.10879v1)

Published 21 Feb 2023 in cs.CL

Abstract: Fine-tuning a LLM on a new domain is standard practice for domain adaptation. However, it can be infeasible when it comes to modern large-scale LLMs such as GPT-3, which can only be accessed through APIs, making it difficult to access the internal parameters of the model. In this paper, we propose $k$NN-Adapter, a method to effectively adapt these black-box LLMs to a new domain. The $k$NN-Adapter builds on top of the retrieval-augmented LLM, and adaptively learns to interpolate the output of the LLM with retrieval results from a datastore consisting of the target domain data. Our experiments on four different domains demonstrate that $k$NN-Adapter significantly improves perplexity, and works particularly well in settings with limited access to LLMs. Additionally, we show that $k$NN-Adapter is more effective than fine-tuning when the amount of training data is limited. We also release a dataset to encourage further study.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Yangsibo Huang (40 papers)
  2. Daogao Liu (34 papers)
  3. Zexuan Zhong (17 papers)
  4. Weijia Shi (55 papers)
  5. Yin Tat Lee (102 papers)
Citations (12)
Youtube Logo Streamline Icon: https://streamlinehq.com