Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Discriminative Language Model as Semantic Consistency Scorer for Prompt-based Few-Shot Text Classification (2210.12763v1)

Published 23 Oct 2022 in cs.CL, cs.AI, and cs.LG

Abstract: This paper proposes a novel prompt-based finetuning method (called DLM-SCS) for few-shot text classification by utilizing the discriminative LLM ELECTRA that is pretrained to distinguish whether a token is original or generated. The underlying idea is that the prompt instantiated with the true label should have higher semantic consistency score than other prompts with false labels. Since a prompt usually consists of several components (or parts), its semantic consistency can be decomposed accordingly. The semantic consistency of each component is then computed by making use of the pretrained ELECTRA model, without introducing extra parameters. Extensive experiments have shown that our model outperforms several state-of-the-art prompt-based few-shot methods.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (2)
  1. Zhipeng Xie (4 papers)
  2. Yahe Li (1 paper)

Summary

We haven't generated a summary for this paper yet.