Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

TriRE: A Multi-Mechanism Learning Paradigm for Continual Knowledge Retention and Promotion (2310.08217v1)

Published 12 Oct 2023 in cs.AI, cs.CV, and cs.LG

Abstract: Continual learning (CL) has remained a persistent challenge for deep neural networks due to catastrophic forgetting (CF) of previously learned tasks. Several techniques such as weight regularization, experience rehearsal, and parameter isolation have been proposed to alleviate CF. Despite their relative success, these research directions have predominantly remained orthogonal and suffer from several shortcomings, while missing out on the advantages of competing strategies. On the contrary, the brain continually learns, accommodates, and transfers knowledge across tasks by simultaneously leveraging several neurophysiological processes, including neurogenesis, active forgetting, neuromodulation, metaplasticity, experience rehearsal, and context-dependent gating, rarely resulting in CF. Inspired by how the brain exploits multiple mechanisms concurrently, we propose TriRE, a novel CL paradigm that encompasses retaining the most prominent neurons for each task, revising and solidifying the extracted knowledge of current and past tasks, and actively promoting less active neurons for subsequent tasks through rewinding and relearning. Across CL settings, TriRE significantly reduces task interference and surpasses different CL approaches considered in isolation.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (57)
  1. “Sparsity and heterogeneous dropout for continual learning in the null space of neural activations” In Conference on Lifelong Learning Agents, 2022, pp. 617–628 PMLR
  2. Wickliffe C Abraham and Anthony Robins “Memory retention–the synaptic stability versus plasticity dilemma” In Trends in neurosciences 28.2 Elsevier, 2005, pp. 73–78
  3. James B Aimone, Janet Wiles and Fred H Gage “Potential role for adult neurogenesis in the encoding of time in new memories” In Nature neuroscience 9.6 Nature Publishing Group US New York, 2006, pp. 723–727
  4. Rahaf Aljundi, Punarjay Chakravarty and Tinne Tuytelaars “Expert gate: Lifelong learning with a network of experts” In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2017, pp. 3366–3375
  5. “Memory aware synapses: Learning what (not) to forget” In Proceedings of the European conference on computer vision (ECCV), 2018, pp. 139–154
  6. Elahe Arani, Fahad Sarfraz and Bahram Zonooz “Learning Fast, Learning Slow: A General Continual Learning Method based on Complementary Learning System” In International Conference on Learning Representations, 2021
  7. Prashant Shivaram Bhat, Bahram Zonooz and Elahe Arani “Consistency is the key to further mitigating catastrophic forgetting in continual learning” In Conference on Lifelong Learning Agents, 2022, pp. 1195–1212 PMLR
  8. Prashant Shivaram Bhat, Bahram Zonooz and Elahe Arani “Task agnostic representation consolidation: a self-supervised based continual learning approach” In Conference on Lifelong Learning Agents, 2022, pp. 390–405 PMLR
  9. “Dark Experience for General Continual Learning: a Strong, Simple Baseline” In Advances in Neural Information Processing Systems 33 Curran Associates, Inc., 2020, pp. 15920–15930
  10. “Rethinking experience replay: a bag of tricks for continual learning” In 2020 25th International Conference on Pattern Recognition (ICPR), 2021, pp. 2180–2187 IEEE
  11. “New Insights on Reducing Abrupt Representation Change in Online Continual Learning” In International Conference on Learning Representations, 2022
  12. Hyuntak Cha, Jaeho Lee and Jinwoo Shin “Co2l: Contrastive continual learning” In Proceedings of the IEEE/CVF International conference on computer vision, 2021, pp. 9516–9525
  13. Diane T Feldman and William C Gordon “The alleviation of short-term retention decrements with reactivation” In Learning and Motivation 10.2 Elsevier, 1979, pp. 198–210
  14. “The Lottery Ticket Hypothesis: Finding Sparse, Trainable Neural Networks” In International Conference on Learning Representations, 2018
  15. “Selective suppression of hippocampal ripples impairs spatial memory” In Nature neuroscience 12.10 Nature Publishing Group US New York, 2009, pp. 1222–1223
  16. Siavash Golkar, Micheal Kagan and Kyunghyun Cho “Continual Learning via Neural Pruning” In Real Neurons & Hidden Units: Future directions at the intersection of neuroscience and artificial intelligence @ NeurIPS 2019, 2019
  17. “Inductive biases for deep learning of higher-level cognition” In Proceedings of the Royal Society A 478.2266 The Royal Society, 2022, pp. 20210068
  18. Mustafa B Gurbuz and Constantine Dovrolis “NISPA: Neuro-Inspired Stability-Plasticity Adaptation for Continual Learning in Sparse Networks” In International Conference on Machine Learning, 2022, pp. 8157–8174 PMLR
  19. Oliver Hardt, Karim Nader and Lynn Nadel “Decay happens: the role of active forgetting in memory” In Trends in cognitive sciences 17.3 Elsevier, 2013, pp. 111–120
  20. “Meta-learning and the neuroscience of learning and memory” In Nature Reviews Neuroscience 18.5 Nature Publishing Group, 2017, pp. Meta–Learning and the Neuroscience of Learning and Memory
  21. “Deep residual learning for image recognition” In Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 770–778
  22. “Learning a unified classifier incrementally via rebalancing” In Proceedings of the IEEE/CVF conference on Computer Vision and Pattern Recognition, 2019, pp. 831–839
  23. “Increasingly packing multiple facial-informatics modules in a unified deep-learning model via lifelong learning” In Proceedings of the 2019 on International Conference on Multimedia Retrieval, 2019, pp. 339–343
  24. Diederik P Kingma and Jimmy Ba “Adam: A method for stochastic optimization” In arXiv preprint arXiv:1412.6980, 2014
  25. “Overcoming catastrophic forgetting in neural networks” In Proceedings of the national academy of sciences 114.13 National Acad Sciences, 2017, pp. 3521–3526
  26. “Biological underpinnings for lifelong learning machines” In Nature Machine Intelligence 4.3 Nature Publishing Group UK London, 2022, pp. 196–210
  27. Jesse J Langille and Richard E Brown “The synaptic theory of memory: a historical survey and reconciliation of recent opposition” In Frontiers in systems neuroscience 12 Frontiers Media SA, 2018, pp. 52
  28. Timothée Lesort, Andrei Stoian and David Filliat “Regularization shortcomings for continual learning” In arXiv preprint arXiv:1912.03049, 2019
  29. Penelope A Lewis and Simon J Durrant “The role of active forgetting in learning and memory” In Frontiers in psychology 9 Frontiers, 2018, pp. 1314
  30. “Learning without forgetting” In IEEE transactions on pattern analysis and machine intelligence 40.12 IEEE, 2017, pp. 2935–2947
  31. “Forgetting generates a novel state that is reactivatable” In Science advances 8.6 American Association for the Advancement of Science, 2022, pp. eabi9071
  32. “Packnet: Adding multiple tasks to a single network by iterative pruning” In Proceedings of the IEEE conference on Computer Vision and Pattern Recognition, 2018, pp. 7765–7773
  33. “Cellular, synaptic and network effects of neuromodulation” In Neural Networks 15.4-6 Elsevier, 2002, pp. 479–493
  34. Nicolas Y Masse, Gregory D Grant and David J Freedman “Alleviating catastrophic forgetting using context-dependent gating and synaptic stabilization” In Proceedings of the National Academy of Sciences 115.44 National Acad Sciences, 2018, pp. E10467–E10475
  35. Michael McCloskey and Neal J Cohen “Catastrophic interference in connectionist networks: The sequential learning problem” In Psychology of learning and motivation 24 Elsevier, 1989, pp. 109–165
  36. Martial Mermillod, Aurélia Bugaiska and Patrick Bonin “The stability-plasticity dilemma: Investigating the continuum from catastrophic forgetting to age-limited learning effects” In Frontiers in psychology 4 Frontiers Media SA, 2013, pp. 504
  37. Vernon B Mountcastle “The columnar organization of the neocortex” In Brain 80.2 Oxford University Press, 1957, pp. 214–241
  38. “Continual lifelong learning with neural networks: A review” In Neural Networks 113 Elsevier, 2019, pp. 54–71
  39. Gyeong-Moon Park, Sahng-Min Yoo and Jong-Hwan Kim “Convolutional neural network with developmental memory for continual learning” In IEEE Transactions on Neural Networks and Learning Systems 32.6 IEEE, 2020, pp. 2691–2705
  40. “Maintaining memories by reactivation” Motor systems / Neurobiology of behaviour In Current Opinion in Neurobiology 17.6, 2007, pp. 698–703
  41. Roger Ratcliff “Connectionist models of recognition memory: constraints imposed by learning and forgetting functions.” In Psychological review 97.2 American Psychological Association, 1990, pp. 285
  42. Anthony Robins “Catastrophic forgetting, rehearsal and pseudorehearsal” In Connection Science 7.2 Taylor & Francis, 1995, pp. 123–146
  43. “Progressive Neural Networks” arXiv, 2016
  44. Fahad Sarfraz, Elahe Arani and Bahram Zonooz “Synergy between synaptic consolidation and experience replay for general continual learning” In Conference on Lifelong Learning Agents, 2022, pp. 920–936 PMLR
  45. “Progress & compress: A scalable framework for continual learning” In International Conference on Machine Learning, 2018, pp. 4528–4537 PMLR
  46. “Use it or lose it: how neurogenesis keeps the brain fit for learning” In Behavioural brain research 227.2 Elsevier, 2012, pp. 450–458
  47. Norman E Spear “Retrieval of memory in animals.” In Psychological Review 80.3 American Psychological Association, 1973, pp. 163
  48. Benjamin C Storm and Elizabeth L Bjork “Active forgetting of episodic memories: Is it time to rethink the ‘retrieval failure’ theory of forgetting?” In British Journal of Psychology 110.3 Wiley Online Library, 2019, pp. 549–571
  49. “GCR: Gradient Coreset Based Replay Buffer Selection For Continual Learning” In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 99–108
  50. “Neural modularity helps organisms evolve to learn new skills without forgetting old skills” In PLoS computational biology 13.2 Public Library of Science, 2017, pp. e1005354
  51. “Foster: Feature boosting and compression for class-incremental learning” In European conference on computer vision, 2022, pp. 398–414 Springer
  52. “Continual learning through retrieval and imagination” In Proceedings of the AAAI Conference on Artificial Intelligence 36.8, 2022, pp. 8594–8602
  53. “SparCL: Sparse Continual Learning on the Edge” In Advances in Neural Information Processing Systems
  54. John T Wixted “The psychology and neuroscience of forgetting” In Annu. Rev. Psychol. 55 Annual Reviews, 2004, pp. 235–269
  55. “Lifelong Learning with Dynamically Expandable Networks” In International Conference on Learning Representations
  56. Friedemann Zenke, Ben Poole and Surya Ganguli “Continual learning through synaptic intelligence” In International conference on machine learning, 2017, pp. 3987–3995 PMLR
  57. “A Model or 603 Exemplars: Towards Memory-Efficient Class-Incremental Learning” In The Eleventh International Conference on Learning Representations, 2022
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Preetha Vijayan (1 paper)
  2. Prashant Bhat (9 papers)
  3. Elahe Arani (59 papers)
  4. Bahram Zonooz (54 papers)
Citations (3)