What Will My Model Forget? Forecasting Forgotten Examples in Language Model Refinement (2402.01865v3)
Abstract: LLMs deployed in the wild make errors. However, simply updating the model with the corrected error instances causes catastrophic forgetting -- the updated model makes errors on instances learned during the instruction tuning or upstream training phase. Randomly replaying upstream data yields unsatisfactory performance and often comes with high variance and poor controllability. To this end, we try to forecast upstream examples that will be forgotten due to a model update for improved controllability of the replay process and interpretability. We train forecasting models given a collection of online learned examples and corresponding forgotten upstream pre-training examples. We propose a partially interpretable forecasting model based on the observation that changes in pre-softmax logit scores of pretraining examples resemble that of online learned examples, which performs decently on BART but fails on T5 models. We further show a black-box classifier based on inner products of example representations achieves better forecasting performance over a series of setups. Finally, we show that we reduce forgetting of upstream pretraining examples by replaying examples that are forecasted to be forgotten, demonstrating the practical utility of forecasting example forgetting.
- Online continual learning with maximal interfered retrieval. In Advances in Neural Information Processing Systems, 2019a.
- Gradient based sample selection for online continual learning. In Neural Information Processing Systems, 2019b.
- Promptsource: An integrated development environment and repository for natural language prompts. In Annual Meeting of the Association for Computational Linguistics: System Demonstrations, 2022.
- Dark experience for general continual learning: a strong, simple baseline. Advances in neural information processing systems, 2020a.
- Rethinking experience replay: a bag of tricks for continual learning. 2020 25th International Conference on Pattern Recognition (ICPR), 2020b.
- Efficient lifelong learning with a-GEM. In International Conference on Learning Representations, 2019.
- Scaling instruction-finetuned language models. arXiv preprint arXiv:2210.11416, 2022.
- Editing factual knowledge in language models. In Conference on Empirical Methods in Natural Language Processing, 2021.
- Episodic memory in lifelong language learning. Advances in Neural Information Processing Systems, 2019.
- A theoretical analysis of catastrophic forgetting through the ntk overlap matrix. In International Conference on Artificial Intelligence and Statistics, 2020.
- How catastrophic can catastrophic forgetting be in linear regression? In Conference on Learning Theory, pp. 4028–4079. PMLR, 2022.
- Aging with grace: Lifelong model editing with discrete key-value adaptors. ArXiv, abs/2211.11031, 2022.
- Measuring massive multitask language understanding. In International Conference on Learning Representations, 2020.
- Lora: Low-rank adaptation of large language models. In International Conference on Learning Representations, 2021.
- Transformer-patcher: One mistake worth one neuron. In International Conference on Learning Representations, 2023.
- Neural tangent kernel: Convergence and generalization in neural networks. Advances in neural information processing systems, 2018.
- Measuring forgetting of memorized training examples. In The Eleventh International Conference on Learning Representations, 2023.
- Towards continual knowledge learning of language models. ArXiv, abs/2110.03215, 2021.
- Lifelong pretraining: Continually adapting language models to emerging corpora. In North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 2022.
- Learning curves for continual learning in neural networks: Self-knowledge transfer and forgetting. In International Conference on Learning Representations, 2021.
- Wide neural networks of any depth evolve as linear models under gradient descent. Advances in neural information processing systems, 2019.
- Bart: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. In Annual Meeting of the Association for Computational Linguistics, 2019.
- Unsupervised cross-task generalization via retrieval augmentation. Advances in Neural Information Processing Systems, 2022a.
- On continual model refinement in out-of-distribution data streams. In Annual Meeting of the Association for Computational Linguistics, 2022b.
- Gradient episodic memory for continual learning. In Neural Information Processing Systems, 2017.
- Characterizing datapoints via second-split forgetting. Advances in Neural Information Processing Systems, 2022.
- Locating and editing factual associations in gpt. In Neural Information Processing Systems, 2022.
- Fast model editing at scale. In International Conference on Learning Representations, 2021.
- Fast finite width neural tangent kernel. In International Conference on Machine Learning, 2022.
- Can LMs learn new entities from descriptions? challenges in propagating injected knowledge. In Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2023.
- OpenAI. Gpt-4 technical report. ArXiv, abs/2303.08774, 2023.
- Raffel, C. Building machine learning models like open source software. Communications of the ACM, 2023.
- Anatomy of catastrophic forgetting: Hidden representations and task semantics. In International Conference on Learning Representations, 2020.
- Robins, A. Catastrophic forgetting, rehearsal and pseudorehearsal. Connection Science, 7(2):123–146, 1995.
- Can BERT refrain from forgetting on sequential tasks? a probing study. In The Eleventh International Conference on Learning Representations, 2023.
- Memorization without overfitting: Analyzing the training dynamics of large language models. Advances in Neural Information Processing Systems, pp. 38274–38290, 2022.
- An empirical study of example forgetting during deep neural network learning. In International Conference on Learning Representations, 2018.
- Pretrained language model in continual learning: A comparative study. In International Conference on Learning Representations, 2021.
- Refining language models with compositional explanations. In Neural Information Processing Systems, 2021.
- Editing large language models: Problems, methods, and opportunities. arXiv preprint arXiv:2305.13172, 2023.
- Online coreset selection for rehearsal-based continual learning. In International Conference on Learning Representations, 2022.
- Plug-and-play knowledge injection for pre-trained language models. In Annual Meeting of the Association for Computational Linguistics, 2023.