Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

LiST: Lite Prompted Self-training Makes Parameter-Efficient Few-shot Learners (2110.06274v2)

Published 12 Oct 2021 in cs.CL

Abstract: We present a new method LiST is short for Lite Prompted Self-Training for parameter-efficient fine-tuning of large pre-trained LLMs (PLMs) for few-shot learning. LiST improves over recent methods that adopt prompt-based fine-tuning (FN) using two key techniques. The first is the use of self-training to leverage large amounts of unlabeled data for prompt-based FN in few-shot settings. We use self-training in conjunction with meta-learning for re-weighting noisy pseudo-prompt labels. Self-training is expensive as it requires updating all the model parameters repetitively. Therefore, we use a second technique for light-weight fine-tuning where we introduce a small number of task-specific parameters that are fine-tuned during self-training while keeping the PLM encoder frozen. Our experiments show that LiST can effectively leverage unlabeled data to improve the model performance for few-shot learning. Additionally, the fine-tuning is efficient as it only updates a small percentage of parameters and the overall model footprint is reduced since several tasks can share a common PLM encoder as backbone. A comprehensive study on six NLU tasks demonstrate LiST to improve by 35% over classic fine-tuning and 6% over prompt-based FN with 96% reduction in number of trainable parameters when fine-tuned with no more than 30 labeled examples from each task. With only 14M tunable parameters, LiST outperforms GPT-3 in-context learning by 33% on few-shot NLU tasks.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Yaqing Wang (59 papers)
  2. Subhabrata Mukherjee (59 papers)
  3. Xiaodong Liu (162 papers)
  4. Jing Gao (98 papers)
  5. Ahmed Hassan Awadallah (50 papers)
  6. Jianfeng Gao (344 papers)
Citations (10)