Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

POUF: Prompt-oriented unsupervised fine-tuning for large pre-trained models (2305.00350v1)

Published 29 Apr 2023 in cs.LG, cs.AI, cs.CL, cs.CV, and stat.ML

Abstract: Through prompting, large-scale pre-trained models have become more expressive and powerful, gaining significant attention in recent years. Though these big models have zero-shot capabilities, in general, labeled data are still required to adapt them to downstream tasks. To overcome this critical limitation, we propose an unsupervised fine-tuning framework to directly fine-tune the model or prompt on the unlabeled target data. We demonstrate how to apply our method to both language-augmented vision and masked-LLMs by aligning the discrete distributions extracted from the prompts and target data. To verify our approach's applicability, we conduct extensive experiments on image classification, sentiment analysis, and natural language inference tasks. Across 13 image-related tasks and 15 language-related ones, the proposed approach achieves consistent improvements over the baselines.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Korawat Tanwisuth (7 papers)
  2. Shujian Zhang (28 papers)
  3. Huangjie Zheng (34 papers)
  4. Pengcheng He (60 papers)
  5. Mingyuan Zhou (161 papers)
Citations (23)