Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Residual Prompt Tuning: Improving Prompt Tuning with Residual Reparameterization (2305.03937v1)

Published 6 May 2023 in cs.CL and cs.AI

Abstract: Prompt tuning is one of the successful approaches for parameter-efficient tuning of pre-trained LLMs. Despite being arguably the most parameter-efficient (tuned soft prompts constitute <0.1% of total parameters), it typically performs worse than other efficient tuning methods and is quite sensitive to hyper-parameters. In this work, we introduce Residual Prompt Tuning - a simple and efficient method that significantly improves the performance and stability of prompt tuning. We propose to reparameterize soft prompt embeddings using a shallow network with a residual connection. Our experiments show that Residual Prompt Tuning significantly outperforms prompt tuning on SuperGLUE benchmark. Notably, our method reaches +7 points improvement over prompt tuning with T5-Base and allows to reduce the prompt length by 10x without hurting performance. In addition, we show that our approach is robust to the choice of learning rate and prompt initialization, and is effective in few-shot settings.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Anastasia Razdaibiedina (7 papers)
  2. Yuning Mao (34 papers)
  3. Rui Hou (56 papers)
  4. Madian Khabsa (38 papers)
  5. Mike Lewis (78 papers)
  6. Jimmy Ba (55 papers)
  7. Amjad Almahairi (19 papers)
Citations (30)