Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Unsupervised Calibration through Prior Adaptation for Text Classification using Large Language Models (2307.06713v3)

Published 13 Jul 2023 in cs.CL and cs.LG

Abstract: A wide variety of natural language tasks are currently being addressed with large-scale LLMs. These models are usually trained with a very large amount of unsupervised text data and adapted to perform a downstream natural language task using methods like fine-tuning, calibration or in-context learning. In this work, we propose an approach to adapt the prior class distribution to perform text classification tasks without the need for labelled samples and only few in-domain sample queries. The proposed approach treats the LLM as a black box, adding a stage where the model posteriors are calibrated to the task. Results show that these methods outperform the un-adapted model for different number of training shots in the prompt and a previous approach were calibration is performed without using any adaptation data.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Lautaro Estienne (4 papers)
  2. Luciana Ferrer (33 papers)
  3. Pablo Piantanida (129 papers)
  4. Matías Vera (4 papers)
Citations (1)