Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Prompt Tuning for Discriminative Pre-trained Language Models (2205.11166v1)

Published 23 May 2022 in cs.CL and cs.AI

Abstract: Recent works have shown promising results of prompt tuning in stimulating pre-trained LLMs (PLMs) for NLP tasks. However, to the best of our knowledge, existing works focus on prompt-tuning generative PLMs that are pre-trained to generate target tokens, such as BERT. It is still unknown whether and how discriminative PLMs, e.g., ELECTRA, can be effectively prompt-tuned. In this work, we present DPT, the first prompt tuning framework for discriminative PLMs, which reformulates NLP tasks into a discriminative LLMing problem. Comprehensive experiments on text classification and question answering show that, compared with vanilla fine-tuning, DPT achieves significantly higher performance, and also prevents the unstable problem in tuning large PLMs in both full-set and low-resource settings. The source code and experiment details of this paper can be obtained from https://github.com/thunlp/DPT.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (9)
  1. Yuan Yao (292 papers)
  2. Bowen Dong (27 papers)
  3. Ao Zhang (45 papers)
  4. Zhengyan Zhang (46 papers)
  5. Ruobing Xie (97 papers)
  6. Zhiyuan Liu (433 papers)
  7. Leyu Lin (43 papers)
  8. Maosong Sun (337 papers)
  9. Jianyong Wang (38 papers)
Citations (31)
Github Logo Streamline Icon: https://streamlinehq.com

GitHub

  1. GitHub - thunlp/DPT (13 stars)