Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
60 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Progressive Prompts: Continual Learning for Language Models (2301.12314v1)

Published 29 Jan 2023 in cs.CL, cs.AI, and cs.LG

Abstract: We introduce Progressive Prompts - a simple and efficient approach for continual learning in LLMs. Our method allows forward transfer and resists catastrophic forgetting, without relying on data replay or a large number of task-specific parameters. Progressive Prompts learns a new soft prompt for each task and sequentially concatenates it with the previously learned prompts, while keeping the base model frozen. Experiments on standard continual learning benchmarks show that our approach outperforms state-of-the-art methods, with an improvement >20% in average test accuracy over the previous best-preforming method on T5 model. We also explore a more challenging continual learning setup with longer sequences of tasks and show that Progressive Prompts significantly outperforms prior methods.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Anastasia Razdaibiedina (7 papers)
  2. Yuning Mao (34 papers)
  3. Rui Hou (56 papers)
  4. Madian Khabsa (38 papers)
  5. Mike Lewis (78 papers)
  6. Amjad Almahairi (19 papers)
Citations (101)
X Twitter Logo Streamline Icon: https://streamlinehq.com