Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
169 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Dual Cognitive Architecture: Incorporating Biases and Multi-Memory Systems for Lifelong Learning (2310.11341v1)

Published 17 Oct 2023 in cs.CV, cs.AI, and cs.LG

Abstract: Artificial neural networks (ANNs) exhibit a narrow scope of expertise on stationary independent data. However, the data in the real world is continuous and dynamic, and ANNs must adapt to novel scenarios while also retaining the learned knowledge to become lifelong learners. The ability of humans to excel at these tasks can be attributed to multiple factors ranging from cognitive computational structures, cognitive biases, and the multi-memory systems in the brain. We incorporate key concepts from each of these to design a novel framework, Dual Cognitive Architecture (DUCA), which includes multiple sub-systems, implicit and explicit knowledge representation dichotomy, inductive bias, and a multi-memory system. The inductive bias learner within DUCA is instrumental in encoding shape information, effectively countering the tendency of ANNs to learn local textures. Simultaneously, the inclusion of a semantic memory submodule facilitates the gradual consolidation of knowledge, replicating the dynamics observed in fast and slow learning systems, reminiscent of the principles underpinning the complementary learning system in human cognition. DUCA shows improvement across different settings and datasets, and it also exhibits reduced task recency bias, without the need for extra information. To further test the versatility of lifelong learning methods on a challenging distribution shift, we introduce a novel domain-incremental dataset DN4IL. In addition to improving performance on existing benchmarks, DUCA also demonstrates superior performance on this complex dataset.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (28)
  1. Learning fast, learning slow: A general continual learning method based on complementary learning system. In International Conference on Learning Representations, 2022. URL https://openreview.net/forum?id=uxxFrDwrE7Y.
  2. Consistency is the key to further mitigating catastrophic forgetting in continual learning. In Conference on Lifelong Learning Agents, pp.  1195–1212. PMLR, 2022.
  3. Dark experience for general continual learning: a strong, simple baseline. Advances in neural information processing systems, 33:15920–15930, 2020.
  4. New insights on reducing abrupt representation change in online continual learning. In International Conference on Learning Representations, 2021.
  5. Co2l: Contrastive continual learning. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp.  9516–9525, 2021.
  6. Dual-process theories in social psychology. Guilford Press, 1999.
  7. François Chollet. On the measure of intelligence. arXiv preprint arXiv:1911.01547, 2019.
  8. On the canny edge detector. Pattern Recognition, 34(3):721–725, 2001.
  9. Towards robust evaluations of continual learning. arXiv preprint arXiv:1805.09733, 2018.
  10. Imagenet-trained CNNs are biased towards texture; increasing shape bias improves accuracy and robustness. In International Conference on Learning Representations, 2019. URL https://openreview.net/forum?id=Bygh9j09KX.
  11. Inbiased: Inductive bias distillation to improve generalization and robustness through shape-awareness. In Conference on Lifelong Learning Agents, pp.  1026–1042. PMLR, 2022.
  12. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp.  770–778, 2016.
  13. On the link between conscious function and general intelligence in humans and machines. arXiv preprint arXiv:2204.05133, 2022.
  14. Learning multiple layers of features from tiny images. 2009.
  15. What learning systems do intelligent agents need? complementary learning systems theory updated. Trends in cognitive sciences, 20(7):512–534, 2016.
  16. Gradient-based learning applied to document recognition. Proceedings of the IEEE, 86(11):2278–2324, 1998.
  17. Catastrophic interference in connectionist networks: The sequential learning problem. In Psychology of learning and motivation, volume 24, pp. 109–165. Elsevier, 1989.
  18. Generalized class incremental learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, pp.  240–241, 2020.
  19. Continual lifelong learning with neural networks: A review. Neural Networks, 113:54–71, 2019.
  20. The book of why: the new science of cause and effect. Basic books, 2018.
  21. Moment matching for multi-source domain adaptation. In Proceedings of the IEEE/CVF international conference on computer vision, pp.  1406–1415, 2019.
  22. Learning to learn without forgetting by maximizing transfer and minimizing interference. In International Conference on Learning Representations, 2018.
  23. Synergy between synaptic consolidation and experience replay for general continual learning. In Conference on Lifelong Learning Agents, pp.  920–936. PMLR, 2022.
  24. A 3x3 isotropic gradient operator for image processing. a talk at the Stanford Artificial Project in, pp.  271–272, 1968.
  25. Computational models of consciousness: A taxonomy and some examples., 2007.
  26. Gido M Van de Ven and Andreas S Tolias. Three scenarios for continual learning. arXiv preprint arXiv:1904.07734, 2019.
  27. Jeffrey S Vitter. Random sampling with a reservoir. ACM Transactions on Mathematical Software (TOMS), 11(1):37–57, 1985.
  28. Large scale incremental learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  374–382, 2019.
Citations (2)

Summary

We haven't generated a summary for this paper yet.