Recent Advances of Foundation Language Models-based Continual Learning: A Survey (2405.18653v2)
Abstract: Recently, foundation LLMs (LMs) have marked significant achievements in the domains of NLP and computer vision (CV). Unlike traditional neural network models, foundation LMs obtain a great ability for transfer learning by acquiring rich commonsense knowledge through pre-training on extensive unsupervised datasets with a vast number of parameters. However, they still can not emulate human-like continuous learning due to catastrophic forgetting. Consequently, various continual learning (CL)-based methodologies have been developed to refine LMs, enabling them to adapt to new tasks without forgetting previous knowledge. However, a systematic taxonomy of existing approaches and a comparison of their performance are still lacking, which is the gap that our survey aims to fill. We delve into a comprehensive review, summarization, and classification of the existing literature on CL-based approaches applied to foundation LLMs, such as pre-trained LLMs (PLMs), LLMs and vision-LLMs (VLMs). We divide these studies into offline CL and online CL, which consist of traditional methods, parameter-efficient-based methods, instruction tuning-based methods and continual pre-training methods. Offline CL encompasses domain-incremental learning, task-incremental learning, and class-incremental learning, while online CL is subdivided into hard task boundary and blurry task boundary settings. Additionally, we outline the typical datasets and metrics employed in CL research and provide a detailed analysis of the challenges and future work for LMs-based continual learning.
- Gpt-4 technical report. arXiv preprint arXiv:2303.08774 (2023).
- Rapid Adaptation in Online Continual Learning: Are We Evaluating It Right?. In ICCV. 18852–18861.
- Online Continual Learning with Maximal Interfered Retrieval. In NeurIPS, Vol. 32.
- Training a helpful and harmless assistant with reinforcement learning from human feedback. arXiv preprint arXiv:2204.05862 (2022).
- Online continual learning on a contaminated data stream with blurry task boundaries. In CVPR. 9275–9284.
- Learning to continually learn. arXiv preprint arXiv:2002.09571 (2020).
- A comprehensive study of class incremental learning algorithms for visual tasks. Neural Networks 135 (2021), 38–54.
- Continual Lifelong Learning in Natural Language Processing: A Survey. In COLING. 6523–6541.
- Raad Bin Tareaf. 2017. Tweets Dataset - Top 20 most followed users in Twitter social platform.
- Léonard Blier and Yann Ollivier. 2018. The description length of deep learning models. NeurIPS 31 (2018).
- Large-scale simple question answering with memory networks. arXiv preprint arXiv:1506.02075 (2015).
- MultiWOZ–a large-scale multi-domain wizard-of-oz dataset for task-oriented dialogue modelling. arXiv preprint arXiv:1810.00278 (2018).
- Taskmaster-1: Toward a realistic and diverse dialog dataset. arXiv preprint arXiv:1909.05358 (2019).
- New insights on reducing abrupt representation change in online continual learning. arXiv preprint arXiv:2104.05025 (2021).
- InstructAlign: High-and-Low Resource Language Alignment via Continual Crosslingual Instruction Tuning. In the First Workshop in South East Asian Language Processing. 55–78.
- Online continual learning with natural distribution shifts: An empirical study with visual data. In ICCV. 8281–8290.
- e-snli: Natural language inference with natural language explanations. NeurIPS 31 (2018).
- Generative Multi-modal Models are Good Class-Incremental Learners. arXiv preprint arXiv:2403.18383 (2024).
- Efficient intent detection with dual sentence encoders. arXiv preprint arXiv:2003.04807 (2020).
- Learning to Solve NLP Tasks in an Incremental Number of Languages. In ACL. 837–847.
- Riemannian walk for incremental learning: Understanding forgetting and intransigence. In ECCV. 532–547.
- Efficient Lifelong Learning with A-GEM. In ICLR.
- CoIN: A Benchmark of Continual Instruction tuNing for Multimodel Large Language Model. arXiv preprint arXiv:2403.08350 (2024).
- Mitigating forgetting in online continual learning via instance-aware parameterization. NeurIPS 33 (2020), 17466–17477.
- Lifelong language pretraining with distribution-specialized experts. In ICML. PMLR, 5383–5395.
- Adapting large language models via reading comprehension. arXiv preprint arXiv:2309.09530 (2023).
- Decouple knowledge from paramters for plug-and-play language modeling. In ACL. 14288–14308.
- QuAC: Question Answering in Context. In EMNLP. 2174–2184.
- Aristotelis Chrysakis and Marie-Francine Moens. 2023. Online bias correction for task-free continual learning. ICLR (2023).
- Continual pre-training mitigates forgetting in language and vision. arXiv preprint arXiv:2205.09357 (2022).
- Snips voice platform: an embedded spoken language understanding system for private-by-design voice interfaces. arXiv preprint arXiv:1805.10190 (2018).
- Continual Vision-Language Retrieval via Dynamic Knowledge Rectification. In AAAI, Vol. 38. 11704–11712.
- Probing representation forgetting in supervised and unsupervised continual learning. In CVPR. 16712–16721.
- A continual learning survey: Defying forgetting in classification tasks. TPAMI 44, 7 (2021), 3366–3385.
- Episodic memory in lifelong language learning. NeurIPS 32 (2019).
- CL-MASR: A Continual Learning Benchmark for Multilingual ASR. arXiv preprint arXiv:2310.16931 (2023).
- Lifelonger: A benchmark for continual disease classification. In International Conference on Medical Image Computing and Computer-Assisted Intervention. 314–324.
- A holistic lexicon-based approach to opinion mining. In WSDM. 231–240.
- A survey on in-context learning. arXiv preprint arXiv:2301.00234 (2022).
- Podnet: Pooled outputs distillation for small-tasks incremental learning. In ECCV. 86–102.
- From Static to Dynamic: A Continual Learning Framework for Large Language Models. arXiv preprint arXiv:2310.14248 (2023).
- A Survey of Vision-Language Pre-Trained Models. In IJCAI. 5436–5443.
- Cynthia Dwork. 2006. Differential privacy. In International colloquium on automata, languages, and programming. 1–12.
- Understanding dataset difficulty with V-usable information. In ICML. 5988–6008.
- From MNIST to ImageNet and Back: Benchmarking Continual Curriculum Learning. arXiv preprint arXiv:2303.11076 (2023).
- Online continual learning under extreme memory constraints. In ECCV 2020. 720–735.
- Continual Learning for Task-oriented Dialogue System with Iterative Network Pruning, Expanding and Masking. In ACL. 517–523.
- A study of continual learning under language shift. arXiv preprint arXiv:2311.01200 (2023).
- Oded Goldreich. 1998. Secure multi-party computation. Manuscript. Preliminary version 78, 110 (1998), 1–108.
- Making the v in vqa matter: Elevating the role of image understanding in visual question answering. In CVPR. 6904–6913.
- The iapr tc-12 benchmark: A new evaluation resource for visual information systems. In International workshop ontoImage, Vol. 2.
- Not just selection, but exploration: Online class-incremental continual learning via dual view consistency. In CVPR. 7442–7451.
- Survey on Online Streaming Continual Learning. In IJCAI.
- Online continual learning through mutual information maximization. In ICML. 8109–8126.
- Dealing with Cross-Task Class Discrimination in Online Continual Learning. In CVPR. 11878–11887.
- Semantic parsing for task oriented dialog using hierarchical representations. arXiv preprint arXiv:1810.07942 (2018).
- DEMix Layers: Disentangling Domains for Modular Language Modeling. In NAACL. 5557–5576.
- FewRel: A large-scale supervised few-shot relation classification dataset with state-of-the-art evaluation. arXiv preprint arXiv:1810.10147 (2018).
- Parameter-Efficient Fine-Tuning for Large Models: A Comprehensive Survey. arXiv preprint arXiv:2403.14608 (2024).
- Replay in deep learning: Current approaches and missing biological elements. Neural computation 33, 11 (2021), 2908–2950.
- Incremental learning in online scenario. In CVPR. 13926–13935.
- Ruining He and Julian McAuley. 2016. Ups and downs: Modeling the visual evolution of fashion trends with one-class collaborative filtering. In WWW. 507–517.
- Meta-learning with sparse experience replay for lifelong language learning. arXiv preprint arXiv:2009.04891 (2020).
- Parameter-efficient transfer learning for NLP. In ICML. 2790–2799.
- OntoNotes: The 90% Solution. In NAACL. 57–60.
- Drinking from a firehose: Continual learning with web-scale natural language. TPAMI 45, 5 (2022), 5684–5696.
- Minqing Hu and Bing Liu. 2004. Mining and summarizing customer reviews. In SIGKDD. 168–177.
- Continual Learning for Text Classification with Information Disentanglement Based Regularization. In ACL. 2736–2746.
- Codesearchnet challenge: Evaluating the state of semantic code search. arXiv preprint arXiv:1909.09436 (2019).
- Mapping language to code in programmatic context. arXiv preprint arXiv:1808.09588 (2018).
- Exploring the benefits of training expert language models over instruction tuning. In ICML. 14702–14729.
- TemporalWiki: A Lifelong Benchmark for Training and Evaluating Ever-Evolving Language Models. In EMNLP. 6237–6250.
- Towards Continual Knowledge Learning of Language Models. In ICLR.
- Khurram Javed and Martha White. 2019. Meta-learning representations for continual learning. NeurIPS 32 (2019).
- CLAP4CLIP: Continual Learning with Probabilistic Finetuning for Vision-Language Models. arXiv preprint arXiv:2403.19137 (2024).
- Learn Continually, Generalize Rapidly: Lifelong Knowledge Accumulation for Few-shot Learning. In EMNLP. 714–729.
- Gradient-based editing of memory examples for online task-free continual learning. NeurIPS 34 (2021), 29193–29205.
- Lifelong Pretraining: Continually Adapting Language Models to Emerging Corpora. In NAACL. 4764–4780.
- TriviaQA: A Large Scale Distantly Supervised Challenge Dataset for Reading Comprehension. In ACL. 1601–1611.
- Kushal Kafle and Christopher Kanan. 2017. An analysis of visual question answering algorithms. In ICCV. 1965–1973.
- Continual Training of Language Models for Few-Shot Learning. In EMNLP. 10205–10216.
- Zixuan Ke and Bing Liu. 2022. Continual learning of natural language processing tasks: A survey. arXiv preprint arXiv:2211.12701 (2022).
- Achieving Forgetting Prevention and Knowledge Transfer in Continual Learning. In NeurIPS, Vol. 34. 22443–22456.
- CLASSIC: Continual and Contrastive Learning of Aspect Sentiment Classification Tasks. In EMNLP.
- Continual Pre-training of Language Models. In ICLR.
- Zixuan Ke and Hu Xu. 2021. Adapting BERT for Continual Learning of a Sequence of Aspect Sentiment Classification Tasks. In NAACL.
- Measuring catastrophic forgetting in neural networks. In AAAI, Vol. 32.
- Jacob Devlin Ming-Wei Chang Kenton and Lee Kristina Toutanova. 2019. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In NAACL. 4171–4186.
- Introducing language guidance in prompt-based continual learning. In ICCV. 11463–11473.
- VLM-PL: Advanced Pseudo Labeling approach Class Incremental Object Detection with Vision-Language Model. arXiv preprint arXiv:2403.05346 (2024).
- Task Relation-aware Continual User Representation Learning. In SIGKDD. 1107–1119.
- Overcoming catastrophic forgetting in neural networks. the national academy of sciences 114, 13 (2017), 3521–3526.
- BeGin: Extensive Benchmark Scenarios and An Easy-to-use Framework for Graph Continual Learning. arXiv preprint arXiv:2211.14568 (2022).
- Online continual learning on class incremental blurry task configuration with anytime inference. arXiv preprint arXiv:2110.10031 (2021).
- Online Boundary-Free Continual Learning by Scheduled Data Prior. In ICLR.
- Hdltex: Hierarchical deep learning for text classification. In ICMLA. 364–371.
- Ken Lang. 1995. Newsweeder: Learning to filter netnews. In Machine learning proceedings. 331–339.
- An evaluation dataset for intent classification and out-of-scope prediction. arXiv preprint arXiv:1909.02027 (2019).
- Do pre-trained models benefit equally in continual learning?. In WCACV. 6485–6493.
- Sungjin Lee. 2017. Toward continual learning for conversational agents. arXiv preprint arXiv:1712.09943 (2017).
- Continual learning for robotics: Definition, framework, learning strategies, opportunities and challenges. Information fusion 58 (2020), 52–68.
- BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension. In ACL.
- Overcoming catastrophic forgetting during domain adaptation of seq2seq language generation. In NAACL. 5441–5454.
- Alime assist: An intelligent assistant for creating an innovative e-commerce experience. In CIKM. 2495–2498.
- Continual few-shot intent detection. In COLING. 333–343.
- Visualbert: A simple and performant baseline for vision and language. arXiv preprint arXiv:1908.03557 (2019).
- PCR: Proxy-based Contrastive Replay for Online Class-Incremental Continual Learning. In CVPR. 24246–24255.
- Microsoft coco: Common objects in context. In ECCV. 740–755.
- The clear benchmark: Continual learning on real-world imagery. In NeurIPS.
- Bing Liu. 2020. Learning on the job: Online lifelong and continual learning. In AAAI, Vol. 34. 13544–13549.
- Bing Liu and Sahisnu Mazumder. 2021. Lifelong and continual learning dialogue systems: learning during conversation. In AAAI, Vol. 35. 15058–15063.
- AI Autonomy: Self-initiated Open-world Continual Learning and Adaptation. AI Magazine (2023).
- Visual instruction tuning. NeurIPS 36 (2024).
- Automated rule selection for aspect extraction in opinion mining. In IJCAI.
- Lifelong intent detection via multi-strategy rebalancing. arXiv preprint arXiv:2108.04445 (2021).
- Continual Learning for Sentence Representations Using Conceptors. In NAACL. 3274–3279.
- Class Incremental Learning with Pre-trained Vision-Language Models. arXiv preprint arXiv:2310.20348 (2023).
- Benchmarking natural language understanding services for building conversational agents. In International Workshop on Spoken Dialogue Systems. 165–183.
- P-Tuning: Prompt Tuning Can Be Comparable to Fine-tuning Across Scales and Tasks. In ACL. 61–68.
- Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692 (2019).
- S2ORC: The Semantic Scholar Open Research Corpus. In ACL. 4969–4983.
- David Lopez-Paz and Marc’Aurelio Ranzato. 2017. Gradient episodic memory for continual learning. NeurIPS 30 (2017).
- Codexglue: A machine learning benchmark dataset for code understanding and generation. arXiv preprint arXiv:2102.04664 (2021).
- Exploring models and data for remote sensing image caption generation. IEEE Transactions on Geoscience and Remote Sensing 56, 4 (2017), 2183–2195.
- An Empirical Study of Catastrophic Forgetting in Large Language Models During Continual Fine-tuning. arXiv e-prints (2023).
- EcomGPT-CT: Continual pre-training of e-commerce large language models with semi-structured data. arXiv preprint arXiv:2312.15696 (2023).
- Learning Word Vectors for Sentiment Analysis. In ACL. 142–150.
- Continual learning in task-oriented dialogue systems. arXiv preprint arXiv:2012.15504 (2020).
- Generative Replay Inspired by Hippocampal Memory Indexing for Continual Language Learning. In EACL. 930–942.
- Online continual learning in image classification: An empirical survey. Neurocomputing 469 (2022), 28–51.
- Class-incremental learning: survey and performance evaluation on image classification. TPAMI 45, 5 (2022), 5513–5533.
- Sahisnu Mazumder and Bing Liu. 2024. Lifelong and Continual Learning Dialogue Systems. Springer Nature.
- The natural language decathlon: Multitask learning as question answering. arXiv 2018. arXiv preprint arXiv:1806.08730 ([n. d.]).
- An Empirical Investigation of the Role of Pre-training in Lifelong Learning. (2021).
- Online Continual Learning in Keyword Spotting for Low-Resource Devices via Pooling High-Order Temporal Statistics. arXiv preprint arXiv:2307.12660 (2023).
- Recent advances in natural language processing via large pre-trained language models: A survey. Comput. Surveys 56, 2 (2023), 1–40.
- Large-scale Lifelong Learning of In-context Instructions and How to Tackle It. In ACL. 12573–12589.
- Continual learning for named entity recognition. In AAAI, Vol. 35. 13570–13577.
- Online Class Incremental Learning on Stochastic Blurry Task Boundary via Mask and Visual Prompt Tuning. In ICCV. 11731–11741.
- A Comprehensive Analysis of Adapter Efficiency. In IKDD. 136–154.
- Continual vision-language representation learning with off-diagonal information. In ICML. 26129–26149.
- The E2E dataset: New challenges for end-to-end generation. arXiv preprint arXiv:1706.09254 (2017).
- Lifelong neural predictive coding: Learning cumulatively online without forgetting. NeurIPS 35 (2022), 5867–5881.
- Training language models to follow instructions with human feedback. NeurIPS 35 (2022), 27730–27744.
- Continual lifelong learning with neural networks: A review. Neural networks 113 (2019), 54–71.
- Scalable Language Model with Generalized Continual Learning. In ICLR.
- Adapters: A Unified Library for Parameter-Efficient and Modular Transfer Learning. In EMNLP. 149–160.
- Decouple before interact: Multi-modal prompt learning for continual visual question answering. In ICCV. 2953–2962.
- Chengwei Qin and Shafiq Joty. 2022a. Continual Few-shot Relation Learning via Embedding Space Regularization and Data Augmentation. In ACL. 2776–2789.
- Chengwei Qin and Shafiq Joty. 2022b. LFPT5: A Unified Framework for Lifelong Few-shot Language Learning Based on Prompt Tuning of T5. In ICLR.
- Recyclable Tuning for Continual Pre-training. In ACL. 11403–11426.
- ELLE: Efficient Lifelong Pre-training for Emerging Data. In ACL. 2789–2810.
- Recent advances of continual learning in computer vision: An overview. arXiv preprint arXiv:2109.11369 (2021).
- Learning transferable visual models from natural language supervision. In ICML. PMLR, 8748–8763.
- SQuAD: 100,000+ Questions for Machine Comprehension of Text. In EMNLP. 2383–2392.
- Zero-shot text-to-image generation. In ICML. Pmlr, 8821–8831.
- Towards scalable multi-domain conversational agents: The schema-guided dialogue dataset. In AAAI, Vol. 34. 8689–8696.
- Progressive Prompts: Continual Learning for Language Models. In ICLR.
- iCaRL: Incremental Classifier and Representation Learning. In CVPR.
- Erik F Sang and Fien De Meulder. 2003. Introduction to the CoNLL-2003 shared task: Language-independent named entity recognition. arXiv preprint cs/0306050 (2003).
- Bloom: A 176b-parameter open-access multilingual language model. arXiv preprint arXiv:2211.05100 (2022).
- Cross-lingual transfer learning for multilingual task oriented dialog. arXiv preprint arXiv:1810.13327 (2018).
- Fine-tuned Language Models are Continual Learners. In EMNLP. 6107–6122.
- Fine-tuned language models are continual learners. In EMNLP. 6107–6122.
- Get to the point: Summarization with pointer-generator networks. arXiv preprint arXiv:1704.04368 (2017).
- Continual-learning-as-a-service (claas): On-demand efficient adaptation of predictive models. arXiv preprint arXiv:2206.06957 (2022).
- Continual learning for real-world autonomous systems: Algorithms, challenges and frameworks. Journal of Intelligent & Robotic Systems 105, 1 (2022), 9.
- A progressive model to enable continual learning for semantic slot filling. In EMNLP. 1279–1284.
- Online class-incremental continual learning with adversarial shapley value. In AAAI, Vol. 35. 9630–9638.
- ConPET: Continual Parameter-Efficient Tuning for Large Language Models. arXiv preprint arXiv:2309.14763 (2023).
- LAMOL: LAnguage MOdeling for Lifelong Language Learning. In ICLR. OpenReview.net.
- Ernie 2.0: A continual pre-training framework for language understanding. In AAAI, Vol. 34. 8968–8975.
- Llama: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971 (2023).
- An empirical study on learning bug-fixing patches in the wild via neural machine translation. TOSEM 28, 4 (2019), 1–29.
- Three types of incremental learning. Nature Machine Intelligence 4, 12 (2022), 1185–1197.
- Steven Vander Eeckt et al. 2023. Rehearsal-Free Online Continual Learning for Automatic Speech Recognition. arXiv e-prints (2023), arXiv–2306.
- Prompt augmented generative replay via supervised contrastive learning for lifelong intent detection. In NAACL. 1113–1127.
- Clad: A realistic continual learning benchmark for autonomous driving. Neural Networks 161 (2023), 659–669.
- TL;DR: Mining Reddit to Learn Automatic Summarization. In the Workshop on New Frontiers in Summarization. 59–63.
- Superglue: A stickier benchmark for general-purpose language understanding systems. NeurIPS 32 (2019).
- GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding. In EMNLP. 353–355.
- Mell: Large-scale extensible user intent classification for dialogue systems with meta lifelong learning. In SIGKDD. 3649–3659.
- RVAE-LAMOL: Residual Variational Autoencoder to Enhance Lifelong Language Learning. In IJCNN. IEEE, 1–9.
- Sentence Embedding Alignment for Lifelong Relation Extraction. In NAACL. 796–806.
- Wanderlust: Online continual object detection in the real world. In ICCV. 10829–10838.
- A comprehensive survey of continual learning: Theory, method and application. TPAMI (2024).
- CBA: Improving Online Continual Learning via Continual Bias Adaptor. In ICCV. 19082–19092.
- Large-scale multi-modal pre-trained models: A comprehensive survey. Machine Intelligence Research 20, 4 (2023), 447–482.
- Orthogonal Subspace Learning for Language Model Continual Learning. In EMNLP. 10658–10671.
- TRACE: A Comprehensive Benchmark for Continual Learning in Large Language Models.
- S-prompts learning with pre-trained transformers: An occam’s razor for domain incremental learning. NeurIPS 35 (2022), 5682–5695.
- Super-naturalinstructions: Generalization via declarative instructions on 1600+ nlp tasks. arXiv preprint arXiv:2204.07705 (2022).
- Rehearsal-free Continual Language Learning via Efficient Parameter Isolation. In ACL. 10933–10946.
- Efficient Meta Lifelong-Learning with Limited Memory. In EMNLP. 535–548.
- Online Prototype Learning for Online Continual Learning. In ICCV. 18764–18774.
- Semantically conditioned lstm-based natural language generation for spoken dialogue systems. arXiv preprint arXiv:1508.01745 (2015).
- Domain-Agnostic Neural Architecture for Class Incremental Continual Learning in Document Processing Platform. In ACL. 527–537.
- Continual world: A robotic benchmark for continual reinforcement learning. NeurIPS 34 (2021), 28496–28510.
- Online Continual Knowledge Learning for Language Models. arXiv preprint arXiv:2311.09632 (2023).
- Incremental few-shot text classification with multi-round new classes: Formulation, dataset and system. arXiv preprint arXiv:2104.11882 (2021).
- Efficient continual pre-training for building domain specific large language models. arXiv preprint arXiv:2311.08545 (2023).
- BERT post-training for review reading comprehension and aspect-based sentiment analysis. arXiv preprint arXiv:1904.02232 (2019).
- Self-taught convolutional neural networks for short text clustering. Neural Networks 88 (2017), 22–31.
- Exploring continual learning for code generation models. arXiv preprint arXiv:2307.02435 (2023).
- M6-t: Exploring sparse expert models and beyond. arXiv preprint arXiv:2105.15082 (2021).
- Towards General Purpose Medical AI: Continual Learning Medical Foundation Model. arXiv preprint arXiv:2303.06580 (2023).
- Mitigating forgetting in online continual learning with neuron calibration. NeurIPS 34 (2021), 10260–10272.
- ConTinTin: Continual Learning from Task Instructions. In ACL. 3062–3072.
- Learning and evaluating general linguistic intelligence. arXiv preprint arXiv:1901.11373 (2019).
- Online coreset selection for rehearsal-based continual learning. arXiv preprint arXiv:2106.01085 (2021).
- From image descriptions to visual denotations: New similarity metrics for semantic inference over event descriptions. TACL 2 (2014), 67–78.
- Boosting Continual Learning of Vision-Language Models via Mixture-of-Experts Adapters. In CVPR.
- Parameter-efficient transfer from sequential behaviors for user modeling and recommendation. In SIGIR. 1469–1478.
- One Person, One Model, One World: Learning Continual User Representation without Forgetting. 696–705.
- Defending against neural fake news. NeurIPS 32 (2019).
- A survey on federated learning. Knowledge-Based Systems 216 (2021), 106775.
- Copf: Continual learning human preference through optimal policy fitting. arXiv preprint arXiv:2310.15694 (2023).
- CPPO: Continual Learning for Reinforcement Learning with Human Feedback. In ICLR.
- Peiyan Zhang and Sunghun Kim. 2023. A Survey on Incremental Update for Neural Recommender Systems. arXiv preprint arXiv:2303.02851 (2023).
- Vqacl: A novel visual question answering continual learning setting. In CVPR. 19102–19112.
- Character-level convolutional networks for text classification. NeurIPS 28 (2015).
- A simple but strong baseline for online continual learning: Repeated augmented rehearsal. NeurIPS 35 (2022), 14771–14783.
- Continual Sequence Generation with Adaptive Compositional Modules. In ACL. 3653–3667.
- Reformulating Domain Adaptation of Large Language Models as Adapt-Retrieve-Revise. arXiv preprint arXiv:2310.03328 (2023).
- Citb: A benchmark for continual instruction tuning. arXiv preprint arXiv:2310.14510 (2023).
- A survey of large language models. arXiv preprint arXiv:2303.18223 (2023).
- Prompt Conditioned VAE: Enhancing Generative Replay for Lifelong Learning in Task-Oriented Dialogue. In EMNLP. 11153–11169.
- Preventing zero-shot transfer degradation in continual learning of vision-language models. In ICCV. 19125–19136.
- Seq2sql: Generating structured queries from natural language using reinforcement learning. arXiv preprint arXiv:1709.00103 (2017).
- Deep class-incremental learning: A survey. arXiv preprint arXiv:2302.03648 (2023).
- Learning without forgetting for vision-language models. arXiv preprint arXiv:2305.19270 (2023).
- ChatGPT: potential, prospects, and limitations. Frontiers of Information Technology & Electronic Engineering (2023), 1–6.
- Ctp: Towards vision-language continual pretraining via compatible momentum contrast and topology preservation. In ICCV. 22257–22267.
- Continual Prompt Tuning for Dialog State Tracking. In ACL. 1124–1137.
- Aligning books and movies: Towards story-like visual explanations by watching movies and reading books. In ICCV. 19–27.
- Yutao Yang (3 papers)
- Jie Zhou (687 papers)
- Xuanwen Ding (4 papers)
- Tianyu Huai (4 papers)
- Shunyu Liu (47 papers)
- Qin Chen (57 papers)
- Liang He (202 papers)
- Yuan Xie (188 papers)