Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Input-Tuning: Adapting Unfamiliar Inputs to Frozen Pretrained Models (2203.03131v1)

Published 7 Mar 2022 in cs.CL and cs.AI

Abstract: Recently the prompt-tuning paradigm has attracted significant attention. By only tuning continuous prompts with a frozen pre-trained LLM (PLM), prompt-tuning takes a step towards deploying a shared frozen PLM to serve numerous downstream tasks. Although prompt-tuning shows good performance on certain natural language understanding (NLU) tasks, its effectiveness on natural language generation (NLG) tasks is still under-explored. In this paper, we argue that one of the factors hindering the development of prompt-tuning on NLG tasks is the unfamiliar inputs (i.e., inputs are linguistically different from the pretraining corpus). For example, our preliminary exploration reveals a large performance gap between prompt-tuning and fine-tuning when unfamiliar inputs occur frequently in NLG tasks. This motivates us to propose input-tuning, which fine-tunes both the continuous prompts and the input representations, leading to a more effective way to adapt unfamiliar inputs to frozen PLMs. Our proposed input-tuning is conceptually simple and empirically powerful. Experimental results on seven NLG tasks demonstrate that input-tuning is significantly and consistently better than prompt-tuning. Furthermore, on three of these tasks, input-tuning can achieve a comparable or even better performance than fine-tuning.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (9)
  1. Shengnan An (12 papers)
  2. Yifei Li (92 papers)
  3. Zeqi Lin (25 papers)
  4. Qian Liu (252 papers)
  5. Bei Chen (56 papers)
  6. Qiang Fu (159 papers)
  7. Weizhu Chen (128 papers)
  8. Nanning Zheng (146 papers)
  9. Jian-Guang Lou (69 papers)
Citations (37)