Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Don't Half-listen: Capturing Key-part Information in Continual Instruction Tuning (2403.10056v1)

Published 15 Mar 2024 in cs.CL and cs.AI

Abstract: Instruction tuning for LLMs can drive them to produce results consistent with human goals in specific downstream tasks. However, the process of continual instruction tuning (CIT) for LLMs may bring about the catastrophic forgetting (CF) problem, where previously learned abilities are degraded. Recent methods try to alleviate the CF problem by modifying models or replaying data, which may only remember the surface-level pattern of instructions and get confused on held-out tasks. In this paper, we propose a novel continual instruction tuning method based on Key-part Information Gain (KPIG). Our method computes the information gain on masked parts to dynamically replay data and refine the training objective, which enables LLMs to capture task-aware information relevant to the correct response and alleviate overfitting to general descriptions in instructions. In addition, we propose two metrics, P-score and V-score, to measure the generalization and instruction-following abilities of LLMs. Experiments demonstrate our method achieves superior performance on both seen and held-out tasks.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Yongquan He (9 papers)
  2. Xuancheng Huang (7 papers)
  3. Minghao Tang (5 papers)
  4. Lingxun Meng (3 papers)
  5. Xiang Li (1002 papers)
  6. Wei Lin (207 papers)
  7. Wenyuan Zhang (30 papers)
  8. Yifu Gao (5 papers)
Citations (3)