Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Few-Shot Learning with Siamese Networks and Label Tuning (2203.14655v2)

Published 28 Mar 2022 in cs.CL and cs.LG

Abstract: We study the problem of building text classifiers with little or no training data, commonly known as zero and few-shot text classification. In recent years, an approach based on neural textual entailment models has been found to give strong results on a diverse range of tasks. In this work, we show that with proper pre-training, Siamese Networks that embed texts and labels offer a competitive alternative. These models allow for a large reduction in inference cost: constant in the number of labels rather than linear. Furthermore, we introduce label tuning, a simple and computationally efficient approach that allows to adapt the models in a few-shot setup by only changing the label embeddings. While giving lower performance than model fine-tuning, this approach has the architectural advantage that a single encoder can be shared by many different tasks.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Thomas Müller (83 papers)
  2. Guillermo Pérez-Torró (2 papers)
  3. Marc Franco-Salvador (13 papers)
Citations (38)

Summary

We haven't generated a summary for this paper yet.