Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Zero-Shot Text Classification via Self-Supervised Tuning (2305.11442v2)

Published 19 May 2023 in cs.CL, cs.AI, and cs.LG

Abstract: Existing solutions to zero-shot text classification either conduct prompting with pre-trained LLMs, which is sensitive to the choices of templates, or rely on large-scale annotated data of relevant tasks for meta-tuning. In this work, we propose a new paradigm based on self-supervised learning to solve zero-shot text classification tasks by tuning the LLMs with unlabeled data, called self-supervised tuning. By exploring the inherent structure of free texts, we propose a new learning objective called first sentence prediction to bridge the gap between unlabeled data and text classification tasks. After tuning the model to learn to predict the first sentence in a paragraph based on the rest, the model is able to conduct zero-shot inference on unseen tasks such as topic classification and sentiment analysis. Experimental results show that our model outperforms the state-of-the-art baselines on 7 out of 10 tasks. Moreover, the analysis reveals that our model is less sensitive to the prompt design. Our code and pre-trained models are publicly available at https://github.com/DAMO-NLP-SG/SSTuning .

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Chaoqun Liu (38 papers)
  2. Wenxuan Zhang (75 papers)
  3. Guizhen Chen (11 papers)
  4. Xiaobao Wu (43 papers)
  5. Anh Tuan Luu (69 papers)
  6. Chip Hong Chang (4 papers)
  7. Lidong Bing (144 papers)
Citations (10)

Summary

We haven't generated a summary for this paper yet.

Github Logo Streamline Icon: https://streamlinehq.com