Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Systematic Analysis for Pretrained Language Model Priming for Parameter-Efficient Fine-tuning (2212.01032v2)

Published 2 Dec 2022 in cs.CL and cs.AI

Abstract: Parameter-efficient (PE) methods (like Prompts or Adapters) for adapting pre-trained LLMs (PLM) to downstream tasks have been popular recently. However, hindrances still prevent these methods from reaching their full potential. For example, two significant challenges are few-shot adaptation and cross-task generalization. To tackle these issues, we propose a general PE priming framework to enhance and explore the few-shot adaptation and generalization ability of PE methods. In this framework, PLMs are primed with PE methods for rapidly adapting to various target tasks. To evaluate the generalization ability of these PE methods, we conduct experiments on a few-shot cross-domain benchmark containing 160 diverse NLP tasks. Our experiment not only reveals the best priming strategy but also verifies that priming facilitates the adaptation to target tasks.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Shih-Cheng Huang (17 papers)
  2. Shih-Heng Wang (6 papers)
  3. Min-Han Shih (3 papers)
  4. Saurav Sahay (34 papers)
  5. Hung-yi Lee (325 papers)