Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Multitask Prompt Tuning Enables Parameter-Efficient Transfer Learning (2303.02861v1)

Published 6 Mar 2023 in cs.CL

Abstract: Prompt tuning, in which a base pretrained model is adapted to each task via conditioning on learned prompt vectors, has emerged as a promising approach for efficiently adapting LLMs to multiple downstream tasks. However, existing methods typically learn soft prompt vectors from scratch, and it has not been clear how to exploit the rich cross-task knowledge with prompt vectors in a multitask learning setting. We propose multitask prompt tuning (MPT), which first learns a single transferable prompt by distilling knowledge from multiple task-specific source prompts. We then learn multiplicative low rank updates to this shared prompt to efficiently adapt it to each downstream target task. Extensive experiments on 23 NLP datasets demonstrate that our proposed approach outperforms the state-of-the-art methods, including the full finetuning baseline in some cases, despite only tuning 0.035% as many task-specific parameters.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Zhen Wang (571 papers)
  2. Rameswar Panda (79 papers)
  3. Leonid Karlinsky (79 papers)
  4. Rogerio Feris (105 papers)
  5. Huan Sun (88 papers)
  6. Yoon Kim (92 papers)
Citations (86)