Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Instance-wise Prompt Tuning for Pretrained Language Models (2206.01958v1)

Published 4 Jun 2022 in cs.CL and cs.AI

Abstract: Prompt Learning has recently gained great popularity in bridging the gap between pretraining tasks and various downstream tasks. It freezes Pretrained LLMs (PLMs) and only tunes a few task-related parameters (prompts) for downstream tasks, greatly reducing the cost of tuning giant models. The key enabler of this is the idea of querying PLMs with task-specific knowledge implicated in prompts. This paper reveals a major limitation of existing methods that the indiscriminate prompts for all input data in a task ignore the intrinsic knowledge from input data, resulting in sub-optimal performance. We introduce Instance-wise Prompt Tuning (IPT), the first prompt learning paradigm that injects knowledge from the input data instances to the prompts, thereby providing PLMs with richer and more concrete context information. We devise a series of strategies to produce instance-wise prompts, addressing various concerns like model quality and cost-efficiency. Across multiple tasks and resource settings, IPT significantly outperforms task-based prompt learning methods, and achieves comparable performance to conventional finetuning with only 0.5% - 1.5% of tuned parameters.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (9)
  1. Yuezihan Jiang (7 papers)
  2. Hao Yang (328 papers)
  3. Junyang Lin (99 papers)
  4. Hanyu Zhao (23 papers)
  5. An Yang (32 papers)
  6. Chang Zhou (105 papers)
  7. Hongxia Yang (130 papers)
  8. Zhi Yang (188 papers)
  9. Bin Cui (165 papers)
Citations (6)