TriRE: A Multi-Mechanism Learning Paradigm for Continual Knowledge Retention and Promotion (2310.08217v1)
Abstract: Continual learning (CL) has remained a persistent challenge for deep neural networks due to catastrophic forgetting (CF) of previously learned tasks. Several techniques such as weight regularization, experience rehearsal, and parameter isolation have been proposed to alleviate CF. Despite their relative success, these research directions have predominantly remained orthogonal and suffer from several shortcomings, while missing out on the advantages of competing strategies. On the contrary, the brain continually learns, accommodates, and transfers knowledge across tasks by simultaneously leveraging several neurophysiological processes, including neurogenesis, active forgetting, neuromodulation, metaplasticity, experience rehearsal, and context-dependent gating, rarely resulting in CF. Inspired by how the brain exploits multiple mechanisms concurrently, we propose TriRE, a novel CL paradigm that encompasses retaining the most prominent neurons for each task, revising and solidifying the extracted knowledge of current and past tasks, and actively promoting less active neurons for subsequent tasks through rewinding and relearning. Across CL settings, TriRE significantly reduces task interference and surpasses different CL approaches considered in isolation.
- “Sparsity and heterogeneous dropout for continual learning in the null space of neural activations” In Conference on Lifelong Learning Agents, 2022, pp. 617–628 PMLR
- Wickliffe C Abraham and Anthony Robins “Memory retention–the synaptic stability versus plasticity dilemma” In Trends in neurosciences 28.2 Elsevier, 2005, pp. 73–78
- James B Aimone, Janet Wiles and Fred H Gage “Potential role for adult neurogenesis in the encoding of time in new memories” In Nature neuroscience 9.6 Nature Publishing Group US New York, 2006, pp. 723–727
- Rahaf Aljundi, Punarjay Chakravarty and Tinne Tuytelaars “Expert gate: Lifelong learning with a network of experts” In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2017, pp. 3366–3375
- “Memory aware synapses: Learning what (not) to forget” In Proceedings of the European conference on computer vision (ECCV), 2018, pp. 139–154
- Elahe Arani, Fahad Sarfraz and Bahram Zonooz “Learning Fast, Learning Slow: A General Continual Learning Method based on Complementary Learning System” In International Conference on Learning Representations, 2021
- Prashant Shivaram Bhat, Bahram Zonooz and Elahe Arani “Consistency is the key to further mitigating catastrophic forgetting in continual learning” In Conference on Lifelong Learning Agents, 2022, pp. 1195–1212 PMLR
- Prashant Shivaram Bhat, Bahram Zonooz and Elahe Arani “Task agnostic representation consolidation: a self-supervised based continual learning approach” In Conference on Lifelong Learning Agents, 2022, pp. 390–405 PMLR
- “Dark Experience for General Continual Learning: a Strong, Simple Baseline” In Advances in Neural Information Processing Systems 33 Curran Associates, Inc., 2020, pp. 15920–15930
- “Rethinking experience replay: a bag of tricks for continual learning” In 2020 25th International Conference on Pattern Recognition (ICPR), 2021, pp. 2180–2187 IEEE
- “New Insights on Reducing Abrupt Representation Change in Online Continual Learning” In International Conference on Learning Representations, 2022
- Hyuntak Cha, Jaeho Lee and Jinwoo Shin “Co2l: Contrastive continual learning” In Proceedings of the IEEE/CVF International conference on computer vision, 2021, pp. 9516–9525
- Diane T Feldman and William C Gordon “The alleviation of short-term retention decrements with reactivation” In Learning and Motivation 10.2 Elsevier, 1979, pp. 198–210
- “The Lottery Ticket Hypothesis: Finding Sparse, Trainable Neural Networks” In International Conference on Learning Representations, 2018
- “Selective suppression of hippocampal ripples impairs spatial memory” In Nature neuroscience 12.10 Nature Publishing Group US New York, 2009, pp. 1222–1223
- Siavash Golkar, Micheal Kagan and Kyunghyun Cho “Continual Learning via Neural Pruning” In Real Neurons & Hidden Units: Future directions at the intersection of neuroscience and artificial intelligence @ NeurIPS 2019, 2019
- “Inductive biases for deep learning of higher-level cognition” In Proceedings of the Royal Society A 478.2266 The Royal Society, 2022, pp. 20210068
- Mustafa B Gurbuz and Constantine Dovrolis “NISPA: Neuro-Inspired Stability-Plasticity Adaptation for Continual Learning in Sparse Networks” In International Conference on Machine Learning, 2022, pp. 8157–8174 PMLR
- Oliver Hardt, Karim Nader and Lynn Nadel “Decay happens: the role of active forgetting in memory” In Trends in cognitive sciences 17.3 Elsevier, 2013, pp. 111–120
- “Meta-learning and the neuroscience of learning and memory” In Nature Reviews Neuroscience 18.5 Nature Publishing Group, 2017, pp. Meta–Learning and the Neuroscience of Learning and Memory
- “Deep residual learning for image recognition” In Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 770–778
- “Learning a unified classifier incrementally via rebalancing” In Proceedings of the IEEE/CVF conference on Computer Vision and Pattern Recognition, 2019, pp. 831–839
- “Increasingly packing multiple facial-informatics modules in a unified deep-learning model via lifelong learning” In Proceedings of the 2019 on International Conference on Multimedia Retrieval, 2019, pp. 339–343
- Diederik P Kingma and Jimmy Ba “Adam: A method for stochastic optimization” In arXiv preprint arXiv:1412.6980, 2014
- “Overcoming catastrophic forgetting in neural networks” In Proceedings of the national academy of sciences 114.13 National Acad Sciences, 2017, pp. 3521–3526
- “Biological underpinnings for lifelong learning machines” In Nature Machine Intelligence 4.3 Nature Publishing Group UK London, 2022, pp. 196–210
- Jesse J Langille and Richard E Brown “The synaptic theory of memory: a historical survey and reconciliation of recent opposition” In Frontiers in systems neuroscience 12 Frontiers Media SA, 2018, pp. 52
- Timothée Lesort, Andrei Stoian and David Filliat “Regularization shortcomings for continual learning” In arXiv preprint arXiv:1912.03049, 2019
- Penelope A Lewis and Simon J Durrant “The role of active forgetting in learning and memory” In Frontiers in psychology 9 Frontiers, 2018, pp. 1314
- “Learning without forgetting” In IEEE transactions on pattern analysis and machine intelligence 40.12 IEEE, 2017, pp. 2935–2947
- “Forgetting generates a novel state that is reactivatable” In Science advances 8.6 American Association for the Advancement of Science, 2022, pp. eabi9071
- “Packnet: Adding multiple tasks to a single network by iterative pruning” In Proceedings of the IEEE conference on Computer Vision and Pattern Recognition, 2018, pp. 7765–7773
- “Cellular, synaptic and network effects of neuromodulation” In Neural Networks 15.4-6 Elsevier, 2002, pp. 479–493
- Nicolas Y Masse, Gregory D Grant and David J Freedman “Alleviating catastrophic forgetting using context-dependent gating and synaptic stabilization” In Proceedings of the National Academy of Sciences 115.44 National Acad Sciences, 2018, pp. E10467–E10475
- Michael McCloskey and Neal J Cohen “Catastrophic interference in connectionist networks: The sequential learning problem” In Psychology of learning and motivation 24 Elsevier, 1989, pp. 109–165
- Martial Mermillod, Aurélia Bugaiska and Patrick Bonin “The stability-plasticity dilemma: Investigating the continuum from catastrophic forgetting to age-limited learning effects” In Frontiers in psychology 4 Frontiers Media SA, 2013, pp. 504
- Vernon B Mountcastle “The columnar organization of the neocortex” In Brain 80.2 Oxford University Press, 1957, pp. 214–241
- “Continual lifelong learning with neural networks: A review” In Neural Networks 113 Elsevier, 2019, pp. 54–71
- Gyeong-Moon Park, Sahng-Min Yoo and Jong-Hwan Kim “Convolutional neural network with developmental memory for continual learning” In IEEE Transactions on Neural Networks and Learning Systems 32.6 IEEE, 2020, pp. 2691–2705
- “Maintaining memories by reactivation” Motor systems / Neurobiology of behaviour In Current Opinion in Neurobiology 17.6, 2007, pp. 698–703
- Roger Ratcliff “Connectionist models of recognition memory: constraints imposed by learning and forgetting functions.” In Psychological review 97.2 American Psychological Association, 1990, pp. 285
- Anthony Robins “Catastrophic forgetting, rehearsal and pseudorehearsal” In Connection Science 7.2 Taylor & Francis, 1995, pp. 123–146
- “Progressive Neural Networks” arXiv, 2016
- Fahad Sarfraz, Elahe Arani and Bahram Zonooz “Synergy between synaptic consolidation and experience replay for general continual learning” In Conference on Lifelong Learning Agents, 2022, pp. 920–936 PMLR
- “Progress & compress: A scalable framework for continual learning” In International Conference on Machine Learning, 2018, pp. 4528–4537 PMLR
- “Use it or lose it: how neurogenesis keeps the brain fit for learning” In Behavioural brain research 227.2 Elsevier, 2012, pp. 450–458
- Norman E Spear “Retrieval of memory in animals.” In Psychological Review 80.3 American Psychological Association, 1973, pp. 163
- Benjamin C Storm and Elizabeth L Bjork “Active forgetting of episodic memories: Is it time to rethink the ‘retrieval failure’ theory of forgetting?” In British Journal of Psychology 110.3 Wiley Online Library, 2019, pp. 549–571
- “GCR: Gradient Coreset Based Replay Buffer Selection For Continual Learning” In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 99–108
- “Neural modularity helps organisms evolve to learn new skills without forgetting old skills” In PLoS computational biology 13.2 Public Library of Science, 2017, pp. e1005354
- “Foster: Feature boosting and compression for class-incremental learning” In European conference on computer vision, 2022, pp. 398–414 Springer
- “Continual learning through retrieval and imagination” In Proceedings of the AAAI Conference on Artificial Intelligence 36.8, 2022, pp. 8594–8602
- “SparCL: Sparse Continual Learning on the Edge” In Advances in Neural Information Processing Systems
- John T Wixted “The psychology and neuroscience of forgetting” In Annu. Rev. Psychol. 55 Annual Reviews, 2004, pp. 235–269
- “Lifelong Learning with Dynamically Expandable Networks” In International Conference on Learning Representations
- Friedemann Zenke, Ben Poole and Surya Ganguli “Continual learning through synaptic intelligence” In International conference on machine learning, 2017, pp. 3987–3995 PMLR
- “A Model or 603 Exemplars: Towards Memory-Efficient Class-Incremental Learning” In The Eleventh International Conference on Learning Representations, 2022
- Preetha Vijayan (1 paper)
- Prashant Bhat (9 papers)
- Elahe Arani (59 papers)
- Bahram Zonooz (54 papers)