Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Pre-trained Token-replaced Detection Model as Few-shot Learner (2203.03235v2)

Published 7 Mar 2022 in cs.CL and cs.AI

Abstract: Pre-trained masked LLMs have demonstrated remarkable ability as few-shot learners. In this paper, as an alternative, we propose a novel approach to few-shot learning with pre-trained token-replaced detection models like ELECTRA. In this approach, we reformulate a classification or a regression task as a token-replaced detection problem. Specifically, we first define a template and label description words for each task and put them into the input to form a natural language prompt. Then, we employ the pre-trained token-replaced detection model to predict which label description word is the most original (i.e., least replaced) among all label description words in the prompt. A systematic evaluation on 16 datasets demonstrates that our approach outperforms few-shot learners with pre-trained masked LLMs in both one-sentence and two-sentence learning tasks.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Zicheng Li (2 papers)
  2. Shoushan Li (6 papers)
  3. Guodong Zhou (62 papers)
Citations (8)

Summary

We haven't generated a summary for this paper yet.